var/home/core/zuul-output/0000755000175000017500000000000015072344307014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015072357325015502 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005062054115072357317017712 0ustar rootrootOct 11 03:05:51 crc systemd[1]: Starting Kubernetes Kubelet... Oct 11 03:05:51 crc restorecon[4720]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:51 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:05:52 crc restorecon[4720]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 11 03:05:52 crc kubenswrapper[4754]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.856799 4754 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864238 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864292 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864301 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864310 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864319 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864327 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864337 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864345 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864353 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864361 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864372 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864437 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864447 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864455 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864464 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864474 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864484 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864492 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864500 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864508 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864516 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864524 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864531 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864539 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864547 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864565 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864573 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864581 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864588 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864596 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864604 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864611 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864619 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864627 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864634 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864642 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864650 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864658 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864666 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864674 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864683 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864690 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864698 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864706 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864715 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864723 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864731 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864739 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864750 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864760 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864770 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864779 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864787 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864796 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864806 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864818 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864829 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864839 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864849 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864859 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864870 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864881 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864891 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864901 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864936 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864949 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.864996 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.865007 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.865016 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.865025 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.865033 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865245 4754 flags.go:64] FLAG: --address="0.0.0.0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865268 4754 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865286 4754 flags.go:64] FLAG: --anonymous-auth="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865300 4754 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865315 4754 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865326 4754 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865343 4754 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865370 4754 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865380 4754 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865389 4754 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865399 4754 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865409 4754 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865418 4754 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865427 4754 flags.go:64] FLAG: --cgroup-root="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865436 4754 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865445 4754 flags.go:64] FLAG: --client-ca-file="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865455 4754 flags.go:64] FLAG: --cloud-config="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865463 4754 flags.go:64] FLAG: --cloud-provider="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865473 4754 flags.go:64] FLAG: --cluster-dns="[]" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865485 4754 flags.go:64] FLAG: --cluster-domain="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865495 4754 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865509 4754 flags.go:64] FLAG: --config-dir="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865520 4754 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865532 4754 flags.go:64] FLAG: --container-log-max-files="5" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865546 4754 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865557 4754 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865569 4754 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865583 4754 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865595 4754 flags.go:64] FLAG: --contention-profiling="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865604 4754 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865613 4754 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865623 4754 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865632 4754 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865643 4754 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865652 4754 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865661 4754 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865669 4754 flags.go:64] FLAG: --enable-load-reader="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865679 4754 flags.go:64] FLAG: --enable-server="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865688 4754 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865719 4754 flags.go:64] FLAG: --event-burst="100" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865730 4754 flags.go:64] FLAG: --event-qps="50" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865738 4754 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865748 4754 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865759 4754 flags.go:64] FLAG: --eviction-hard="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865770 4754 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865779 4754 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865788 4754 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865798 4754 flags.go:64] FLAG: --eviction-soft="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865807 4754 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865816 4754 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865825 4754 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865835 4754 flags.go:64] FLAG: --experimental-mounter-path="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865844 4754 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865854 4754 flags.go:64] FLAG: --fail-swap-on="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865864 4754 flags.go:64] FLAG: --feature-gates="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865878 4754 flags.go:64] FLAG: --file-check-frequency="20s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865890 4754 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865901 4754 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865913 4754 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865926 4754 flags.go:64] FLAG: --healthz-port="10248" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865938 4754 flags.go:64] FLAG: --help="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865949 4754 flags.go:64] FLAG: --hostname-override="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865989 4754 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.865999 4754 flags.go:64] FLAG: --http-check-frequency="20s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866009 4754 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866048 4754 flags.go:64] FLAG: --image-credential-provider-config="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866059 4754 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866068 4754 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866077 4754 flags.go:64] FLAG: --image-service-endpoint="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866152 4754 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866163 4754 flags.go:64] FLAG: --kube-api-burst="100" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866173 4754 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866183 4754 flags.go:64] FLAG: --kube-api-qps="50" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866192 4754 flags.go:64] FLAG: --kube-reserved="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866201 4754 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866210 4754 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866220 4754 flags.go:64] FLAG: --kubelet-cgroups="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866228 4754 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866237 4754 flags.go:64] FLAG: --lock-file="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866248 4754 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866258 4754 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866267 4754 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866281 4754 flags.go:64] FLAG: --log-json-split-stream="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866290 4754 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866299 4754 flags.go:64] FLAG: --log-text-split-stream="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866308 4754 flags.go:64] FLAG: --logging-format="text" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866317 4754 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866327 4754 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866336 4754 flags.go:64] FLAG: --manifest-url="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866345 4754 flags.go:64] FLAG: --manifest-url-header="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866356 4754 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866366 4754 flags.go:64] FLAG: --max-open-files="1000000" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866377 4754 flags.go:64] FLAG: --max-pods="110" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866386 4754 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866395 4754 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866404 4754 flags.go:64] FLAG: --memory-manager-policy="None" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866413 4754 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866422 4754 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866432 4754 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866441 4754 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866461 4754 flags.go:64] FLAG: --node-status-max-images="50" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866470 4754 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866479 4754 flags.go:64] FLAG: --oom-score-adj="-999" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866488 4754 flags.go:64] FLAG: --pod-cidr="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866497 4754 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866511 4754 flags.go:64] FLAG: --pod-manifest-path="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866520 4754 flags.go:64] FLAG: --pod-max-pids="-1" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866529 4754 flags.go:64] FLAG: --pods-per-core="0" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866539 4754 flags.go:64] FLAG: --port="10250" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866548 4754 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866558 4754 flags.go:64] FLAG: --provider-id="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866567 4754 flags.go:64] FLAG: --qos-reserved="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866576 4754 flags.go:64] FLAG: --read-only-port="10255" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866585 4754 flags.go:64] FLAG: --register-node="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866594 4754 flags.go:64] FLAG: --register-schedulable="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866605 4754 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866634 4754 flags.go:64] FLAG: --registry-burst="10" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866646 4754 flags.go:64] FLAG: --registry-qps="5" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866657 4754 flags.go:64] FLAG: --reserved-cpus="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866667 4754 flags.go:64] FLAG: --reserved-memory="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866682 4754 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866693 4754 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866704 4754 flags.go:64] FLAG: --rotate-certificates="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866716 4754 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866727 4754 flags.go:64] FLAG: --runonce="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866737 4754 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866746 4754 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866756 4754 flags.go:64] FLAG: --seccomp-default="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866765 4754 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866774 4754 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866784 4754 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866793 4754 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866803 4754 flags.go:64] FLAG: --storage-driver-password="root" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866812 4754 flags.go:64] FLAG: --storage-driver-secure="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866821 4754 flags.go:64] FLAG: --storage-driver-table="stats" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866830 4754 flags.go:64] FLAG: --storage-driver-user="root" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866839 4754 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866848 4754 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866858 4754 flags.go:64] FLAG: --system-cgroups="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866867 4754 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866881 4754 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866890 4754 flags.go:64] FLAG: --tls-cert-file="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866898 4754 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.866999 4754 flags.go:64] FLAG: --tls-min-version="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867009 4754 flags.go:64] FLAG: --tls-private-key-file="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867018 4754 flags.go:64] FLAG: --topology-manager-policy="none" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867027 4754 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867036 4754 flags.go:64] FLAG: --topology-manager-scope="container" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867046 4754 flags.go:64] FLAG: --v="2" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867058 4754 flags.go:64] FLAG: --version="false" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867071 4754 flags.go:64] FLAG: --vmodule="" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867084 4754 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.867094 4754 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867350 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867362 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867372 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867384 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867393 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867401 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867409 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867417 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867425 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867433 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867441 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867449 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867457 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867464 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867472 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867480 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867488 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867496 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867503 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867512 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867520 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867528 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867535 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867546 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867556 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867566 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867574 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867583 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867592 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867600 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867608 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867618 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867629 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867640 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867650 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867660 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867669 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867677 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867685 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867692 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867700 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867708 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867757 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867771 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867782 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867792 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867803 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867812 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867822 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867833 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867842 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867850 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867859 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867866 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867874 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867882 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867890 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867897 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867905 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867913 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867921 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867929 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867937 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867944 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867952 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.867989 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.868000 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.868010 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.868018 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.868027 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.868037 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.868060 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.877401 4754 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.877434 4754 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877503 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877513 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877518 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877523 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877529 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877537 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877542 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877548 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877552 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877556 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877560 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877564 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877568 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877574 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877580 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877585 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877590 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877594 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877599 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877603 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877608 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877612 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877616 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877621 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877625 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877630 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877634 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877639 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877643 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877647 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877652 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877656 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877661 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877665 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877671 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877676 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877680 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877685 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877689 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877693 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877699 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877705 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877710 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877714 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877718 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877723 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877727 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877732 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877736 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877740 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877744 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877749 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877753 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877758 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877762 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877766 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877770 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877774 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877779 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877782 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877786 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877789 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877793 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877797 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877800 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877803 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877807 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877810 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877814 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877817 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877821 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.877828 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877942 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877949 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877955 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877981 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877990 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877995 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.877999 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878004 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878030 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878035 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878041 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878046 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878050 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878054 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878061 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878066 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878071 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878075 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878079 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878083 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878087 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878091 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878095 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878099 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878102 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878106 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878109 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878113 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878117 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878120 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878124 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878127 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878131 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878135 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878139 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878143 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878146 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878150 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878153 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878157 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878160 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878163 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878167 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878171 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878174 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878178 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878181 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878185 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878189 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878193 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878196 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878200 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878204 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878208 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878211 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878215 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878218 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878222 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878225 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878229 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878232 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878236 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878239 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878244 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878249 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878253 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878257 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878261 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878265 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878268 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:05:52 crc kubenswrapper[4754]: W1011 03:05:52.878272 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.878279 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.879660 4754 server.go:940] "Client rotation is on, will bootstrap in background" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.884769 4754 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.884843 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.886272 4754 server.go:997] "Starting client certificate rotation" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.886292 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.886408 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-21 06:23:34.809689709 +0000 UTC Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.886473 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1707h17m41.92321968s for next certificate rotation Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.909250 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.910925 4754 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.928917 4754 log.go:25] "Validated CRI v1 runtime API" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.971139 4754 log.go:25] "Validated CRI v1 image API" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.972903 4754 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.977116 4754 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-11-03-01-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.977167 4754 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.993830 4754 manager.go:217] Machine: {Timestamp:2025-10-11 03:05:52.990912762 +0000 UTC m=+0.549857567 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:abd9917e-dee0-4bd6-812d-a0709f74ee02 BootID:f5c76452-3606-44cd-ab8e-b64535e75615 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:95:e3:e5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:95:e3:e5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:38:84:84 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:4c:90:25 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:16:c9:54 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:fd:6d:18 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:18:9b:45 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:27:85:f8:ce:73 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:de:ee:f4:ff:48 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.994087 4754 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.994228 4754 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.996104 4754 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.996318 4754 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.996365 4754 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.996564 4754 topology_manager.go:138] "Creating topology manager with none policy" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.996577 4754 container_manager_linux.go:303] "Creating device plugin manager" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.997048 4754 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.997079 4754 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.997367 4754 state_mem.go:36] "Initialized new in-memory state store" Oct 11 03:05:52 crc kubenswrapper[4754]: I1011 03:05:52.997457 4754 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.001296 4754 kubelet.go:418] "Attempting to sync node with API server" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.001323 4754 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.001338 4754 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.001351 4754 kubelet.go:324] "Adding apiserver pod source" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.001362 4754 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.006434 4754 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.009467 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.009629 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.009707 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.009812 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.010010 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.011568 4754 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016025 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016103 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016272 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016283 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016297 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016305 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016313 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016326 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016338 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016346 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016361 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.016369 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.017578 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.019015 4754 server.go:1280] "Started kubelet" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.022175 4754 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.022199 4754 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.022761 4754 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.022926 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc systemd[1]: Started Kubernetes Kubelet. Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025286 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025361 4754 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025533 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:48:28.385734456 +0000 UTC Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025615 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2301h42m35.360125493s for next certificate rotation Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025677 4754 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025693 4754 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.025835 4754 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.026101 4754 server.go:460] "Adding debug handlers to kubelet server" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.026396 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.026414 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="200ms" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.026493 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.025325 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.188:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d50e026678105 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-11 03:05:53.018945797 +0000 UTC m=+0.577890592,LastTimestamp:2025-10-11 03:05:53.018945797 +0000 UTC m=+0.577890592,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.026833 4754 factory.go:55] Registering systemd factory Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.028033 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.028548 4754 factory.go:221] Registration of the systemd container factory successfully Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.029171 4754 factory.go:153] Registering CRI-O factory Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.029193 4754 factory.go:221] Registration of the crio container factory successfully Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.029260 4754 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.029284 4754 factory.go:103] Registering Raw factory Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.029299 4754 manager.go:1196] Started watching for new ooms in manager Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.031125 4754 manager.go:319] Starting recovery of all containers Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.033131 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.033200 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.033216 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.033228 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.033239 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.034996 4754 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035094 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035112 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035126 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035144 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035156 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035171 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035183 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035196 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035214 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035227 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035239 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035256 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035270 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035282 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035293 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035308 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035322 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035336 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035349 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035363 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035375 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035390 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035404 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035417 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035432 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035445 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035459 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035473 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035487 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035501 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035513 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035526 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035552 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035565 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035579 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035590 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035603 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035617 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035632 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035645 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035660 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035678 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035691 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035703 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035716 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035730 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035745 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035764 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035779 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035794 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035809 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035824 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035838 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035851 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035866 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035880 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035896 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035908 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035921 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035933 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035945 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035974 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.035989 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036001 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036017 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036033 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036046 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036058 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036072 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036085 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036099 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036113 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036127 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036140 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036155 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036168 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036182 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036195 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036210 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036225 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036239 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036252 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036268 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036347 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036364 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036378 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036392 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036408 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036423 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036438 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036453 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036466 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036477 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036529 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036546 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036561 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036577 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036589 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036601 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036624 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036656 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036672 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036690 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036703 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036717 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036730 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036745 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036758 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036772 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036787 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036799 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036814 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036831 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036843 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036855 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036868 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036899 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036912 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036924 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.036935 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.040467 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.042991 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043083 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043119 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043220 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043253 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043311 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043373 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043393 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043411 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043434 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043451 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043471 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043493 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043551 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043579 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043781 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043823 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043887 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043904 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043924 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.043945 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044080 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044102 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044118 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044151 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044167 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044286 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044313 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044336 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.044786 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045417 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045446 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045468 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045654 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045673 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045688 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045702 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045719 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045732 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045801 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.045821 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046131 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046368 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046474 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046613 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046643 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046794 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046812 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046878 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046943 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046973 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.046997 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047020 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047044 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047064 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047083 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047147 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047247 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047269 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047333 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047356 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047382 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047426 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047446 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047463 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047584 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047612 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047627 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047688 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047705 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047717 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047770 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.047841 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048645 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048669 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048682 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048697 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048712 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048723 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048737 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048749 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048761 4754 reconstruct.go:97] "Volume reconstruction finished" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.048769 4754 reconciler.go:26] "Reconciler: start to sync state" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.056083 4754 manager.go:324] Recovery completed Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.066076 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.067441 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.067478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.067489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.068111 4754 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.068150 4754 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.068169 4754 state_mem.go:36] "Initialized new in-memory state store" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.080167 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.082337 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.082384 4754 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.082413 4754 kubelet.go:2335] "Starting kubelet main sync loop" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.082565 4754 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.083447 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.083500 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.085200 4754 policy_none.go:49] "None policy: Start" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.086465 4754 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.086575 4754 state_mem.go:35] "Initializing new in-memory state store" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.128613 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.143180 4754 manager.go:334] "Starting Device Plugin manager" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.143234 4754 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.143246 4754 server.go:79] "Starting device plugin registration server" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.143768 4754 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.143784 4754 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.144540 4754 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.145098 4754 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.145120 4754 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.153401 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.182832 4754 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.182944 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.185688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.185722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.185732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.185928 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.186414 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.186471 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.187723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.187839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.187771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.187885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.187910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.188061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.188385 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.188568 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.188605 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.190294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.190317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.190377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191382 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191737 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.191833 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192495 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192717 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.192790 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193618 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193913 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193974 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.193985 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.194707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.194744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.194753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.227593 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="400ms" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.244909 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.246398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.246474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.246500 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.246562 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.247541 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.188:6443: connect: connection refused" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.250798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.250870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.250932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251020 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251073 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251141 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251374 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251455 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251509 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251557 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.251584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353348 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353451 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353499 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353602 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353827 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353713 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353910 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353947 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.353998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354039 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354104 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354210 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354220 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.354459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.448678 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.450419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.450455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.450471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.450500 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.450990 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.188:6443: connect: connection refused" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.508985 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.514707 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.534368 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.554916 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.556626 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-54c8fac6759f8d8c4610a435f06cfa7122101482622d1d1cb8c1f55a4cd31d58 WatchSource:0}: Error finding container 54c8fac6759f8d8c4610a435f06cfa7122101482622d1d1cb8c1f55a4cd31d58: Status 404 returned error can't find the container with id 54c8fac6759f8d8c4610a435f06cfa7122101482622d1d1cb8c1f55a4cd31d58 Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.558945 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-de82a1c404f32500d9ba7bbc1bda3e21edb7d72a08380f97a286d7304080aa0d WatchSource:0}: Error finding container de82a1c404f32500d9ba7bbc1bda3e21edb7d72a08380f97a286d7304080aa0d: Status 404 returned error can't find the container with id de82a1c404f32500d9ba7bbc1bda3e21edb7d72a08380f97a286d7304080aa0d Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.560329 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.564252 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-071d4b584fbddd2ba3cf6d06a409a61c7a1b034e154b73d401d22d5205ef2a47 WatchSource:0}: Error finding container 071d4b584fbddd2ba3cf6d06a409a61c7a1b034e154b73d401d22d5205ef2a47: Status 404 returned error can't find the container with id 071d4b584fbddd2ba3cf6d06a409a61c7a1b034e154b73d401d22d5205ef2a47 Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.573063 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-1c80dd59d2e25b8b00c1f46636eae6b0bb7b69975c174bd7c611ade5344c46a8 WatchSource:0}: Error finding container 1c80dd59d2e25b8b00c1f46636eae6b0bb7b69975c174bd7c611ade5344c46a8: Status 404 returned error can't find the container with id 1c80dd59d2e25b8b00c1f46636eae6b0bb7b69975c174bd7c611ade5344c46a8 Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.580669 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-5ef482e5eab5ab7c96ef8b8dadda295f0d111a5a43937f1855203d53303403c2 WatchSource:0}: Error finding container 5ef482e5eab5ab7c96ef8b8dadda295f0d111a5a43937f1855203d53303403c2: Status 404 returned error can't find the container with id 5ef482e5eab5ab7c96ef8b8dadda295f0d111a5a43937f1855203d53303403c2 Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.629099 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="800ms" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.851904 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.853789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.853844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.853859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:53 crc kubenswrapper[4754]: I1011 03:05:53.853890 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.854443 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.188:6443: connect: connection refused" node="crc" Oct 11 03:05:53 crc kubenswrapper[4754]: W1011 03:05:53.891566 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:53 crc kubenswrapper[4754]: E1011 03:05:53.891716 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.023951 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.087759 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ef482e5eab5ab7c96ef8b8dadda295f0d111a5a43937f1855203d53303403c2"} Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.088494 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c80dd59d2e25b8b00c1f46636eae6b0bb7b69975c174bd7c611ade5344c46a8"} Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.089255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"071d4b584fbddd2ba3cf6d06a409a61c7a1b034e154b73d401d22d5205ef2a47"} Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.090238 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de82a1c404f32500d9ba7bbc1bda3e21edb7d72a08380f97a286d7304080aa0d"} Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.090992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"54c8fac6759f8d8c4610a435f06cfa7122101482622d1d1cb8c1f55a4cd31d58"} Oct 11 03:05:54 crc kubenswrapper[4754]: W1011 03:05:54.270467 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:54 crc kubenswrapper[4754]: E1011 03:05:54.270547 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:54 crc kubenswrapper[4754]: E1011 03:05:54.430040 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="1.6s" Oct 11 03:05:54 crc kubenswrapper[4754]: W1011 03:05:54.533927 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:54 crc kubenswrapper[4754]: E1011 03:05:54.534108 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:54 crc kubenswrapper[4754]: W1011 03:05:54.571340 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:54 crc kubenswrapper[4754]: E1011 03:05:54.571501 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.654865 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.656333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.656384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.656404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:54 crc kubenswrapper[4754]: I1011 03:05:54.656443 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:54 crc kubenswrapper[4754]: E1011 03:05:54.657080 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.188:6443: connect: connection refused" node="crc" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.024500 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.094450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.094500 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.094513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.094521 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.094604 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.095378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.095405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.095413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.096818 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad" exitCode=0 Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.096867 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.096946 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.098232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.098265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.098275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.100649 4754 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857" exitCode=0 Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.100724 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.100799 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.101798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.101818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.101826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.103779 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1" exitCode=0 Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.103871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.103921 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105781 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4" exitCode=0 Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105809 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4"} Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.105892 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.106628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.106649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.106658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.111011 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.111849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.111881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:55 crc kubenswrapper[4754]: I1011 03:05:55.111900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.024016 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:56 crc kubenswrapper[4754]: E1011 03:05:56.031507 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="3.2s" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.110786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.110814 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.111578 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.111616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.111633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.113660 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.113703 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.113719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.113706 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.114627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.114659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.114671 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116425 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116439 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116456 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116490 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.116504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117856 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980" exitCode=0 Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117886 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980"} Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.117989 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.118017 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.118824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.118849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.118856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.118988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.119017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.119031 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: W1011 03:05:56.215079 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:56 crc kubenswrapper[4754]: E1011 03:05:56.215169 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.257853 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.259156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.259192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.259203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.259228 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:56 crc kubenswrapper[4754]: E1011 03:05:56.259689 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.188:6443: connect: connection refused" node="crc" Oct 11 03:05:56 crc kubenswrapper[4754]: W1011 03:05:56.501716 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.188:6443: connect: connection refused Oct 11 03:05:56 crc kubenswrapper[4754]: E1011 03:05:56.501823 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.188:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.536547 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.536866 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Oct 11 03:05:56 crc kubenswrapper[4754]: I1011 03:05:56.536933 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.125692 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2" exitCode=0 Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.125828 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.125873 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.125929 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.125989 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.126106 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.126122 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.126686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2"} Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.127941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.128019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.128076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.128112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:57 crc kubenswrapper[4754]: I1011 03:05:57.777111 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.132907 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765"} Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.132973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d"} Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.132992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6"} Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.133005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e"} Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.133007 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.133918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.133994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.134009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:58 crc kubenswrapper[4754]: I1011 03:05:58.971429 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.041309 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.041661 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.043784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.043848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.043863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.141897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589"} Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.142045 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.142056 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.143555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.219742 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.220038 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.221480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.221542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.221560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.460233 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.464150 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.464209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.464222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.464251 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:05:59 crc kubenswrapper[4754]: I1011 03:05:59.674062 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.144088 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.144168 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.145890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.821137 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.821352 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.822672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.822733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.822756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:00 crc kubenswrapper[4754]: I1011 03:06:00.831640 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.146727 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.146863 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148088 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:01 crc kubenswrapper[4754]: I1011 03:06:01.148177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.071693 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.071879 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.073354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.073476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.073504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:03 crc kubenswrapper[4754]: E1011 03:06:03.153686 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.202287 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.202473 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.203653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.203734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:03 crc kubenswrapper[4754]: I1011 03:06:03.203755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:04 crc kubenswrapper[4754]: I1011 03:06:04.505432 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 11 03:06:04 crc kubenswrapper[4754]: I1011 03:06:04.505619 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:04 crc kubenswrapper[4754]: I1011 03:06:04.506855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:04 crc kubenswrapper[4754]: I1011 03:06:04.506884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:04 crc kubenswrapper[4754]: I1011 03:06:04.506893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.071792 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.071887 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 03:06:06 crc kubenswrapper[4754]: W1011 03:06:06.784620 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.784760 4754 trace.go:236] Trace[583922358]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:05:56.782) (total time: 10001ms): Oct 11 03:06:06 crc kubenswrapper[4754]: Trace[583922358]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:06:06.784) Oct 11 03:06:06 crc kubenswrapper[4754]: Trace[583922358]: [10.001973675s] [10.001973675s] END Oct 11 03:06:06 crc kubenswrapper[4754]: E1011 03:06:06.784794 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.887068 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.887173 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.893548 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 11 03:06:06 crc kubenswrapper[4754]: I1011 03:06:06.893590 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 11 03:06:09 crc kubenswrapper[4754]: I1011 03:06:09.226669 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:09 crc kubenswrapper[4754]: I1011 03:06:09.226908 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:09 crc kubenswrapper[4754]: I1011 03:06:09.228506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:09 crc kubenswrapper[4754]: I1011 03:06:09.228735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:09 crc kubenswrapper[4754]: I1011 03:06:09.228878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.245506 4754 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.545445 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.545622 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.546882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.546921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.546940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.550784 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:11 crc kubenswrapper[4754]: E1011 03:06:11.879812 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 11 03:06:11 crc kubenswrapper[4754]: E1011 03:06:11.888790 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.890677 4754 trace.go:236] Trace[865629081]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:06:00.180) (total time: 11709ms): Oct 11 03:06:11 crc kubenswrapper[4754]: Trace[865629081]: ---"Objects listed" error: 11709ms (03:06:11.890) Oct 11 03:06:11 crc kubenswrapper[4754]: Trace[865629081]: [11.709966782s] [11.709966782s] END Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.890726 4754 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.891321 4754 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.895471 4754 trace.go:236] Trace[1917535662]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:05:57.435) (total time: 14460ms): Oct 11 03:06:11 crc kubenswrapper[4754]: Trace[1917535662]: ---"Objects listed" error: 14460ms (03:06:11.895) Oct 11 03:06:11 crc kubenswrapper[4754]: Trace[1917535662]: [14.460321321s] [14.460321321s] END Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.895529 4754 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 11 03:06:11 crc kubenswrapper[4754]: I1011 03:06:11.901797 4754 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.010934 4754 apiserver.go:52] "Watching apiserver" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.016915 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.017216 4754 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.017683 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018107 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.018184 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018109 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018433 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.018493 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018526 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.018525 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.018691 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.020106 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.020392 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.020457 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.020827 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.021145 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.021314 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.021441 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.021573 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.024898 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.027151 4754 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.032427 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.032469 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.062392 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.076683 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.087560 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093019 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093053 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093079 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093106 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093130 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093259 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093287 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093308 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093367 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093391 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093416 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093445 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093467 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093408 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093544 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093437 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.093520 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:12.593485388 +0000 UTC m=+20.152430233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093638 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093532 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093770 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093846 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093875 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093903 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093927 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093950 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.093994 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094040 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094057 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094182 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094260 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094266 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094281 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094278 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094379 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094421 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094457 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094487 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094519 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094583 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094621 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094652 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094318 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094329 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094328 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094497 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094557 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094570 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094666 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094686 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094773 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094791 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094804 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094810 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094885 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094917 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094942 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094985 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095009 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095060 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095084 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095104 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095128 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095148 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095171 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095222 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095246 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095267 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095294 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095315 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095385 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095410 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095432 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095464 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095490 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.094907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095017 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095013 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095155 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095287 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095348 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095414 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095512 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095638 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095533 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095703 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095730 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095743 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095763 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095787 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095811 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095844 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095869 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095886 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095891 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095940 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.095992 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096025 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096073 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096099 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096141 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096152 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096202 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096225 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096235 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096250 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096292 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096306 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096331 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096358 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096376 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096381 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096426 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096451 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096481 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096502 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096537 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096550 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096588 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096690 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096771 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096859 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096887 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097143 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097141 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097365 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097576 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097583 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.096548 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097730 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097780 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097821 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097852 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097882 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097894 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097915 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097954 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098020 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098081 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098117 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098294 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098329 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098362 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098412 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098448 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098502 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098582 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098617 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098651 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098685 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098717 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098753 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099428 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099493 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099530 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099590 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099623 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099656 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099687 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099726 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099763 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099796 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099870 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099905 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099938 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099999 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100070 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100093 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100117 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100141 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100197 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100346 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100376 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100397 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100421 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101501 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101549 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101592 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101659 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101692 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101729 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101763 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101799 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101833 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101899 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101935 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101994 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102089 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102127 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102163 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102187 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102249 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102301 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102338 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102374 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102419 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102453 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102485 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102518 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102553 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102589 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102622 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102654 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.097945 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098053 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098189 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098214 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098236 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098318 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098430 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098614 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098609 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.098765 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.099680 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100647 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100681 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.100722 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101327 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.101831 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102521 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102604 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102671 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103014 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103198 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.102693 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103291 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103317 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103337 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103355 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103380 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103338 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103496 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103635 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103646 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103857 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103874 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103890 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103907 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103922 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.103946 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104018 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104051 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104113 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104133 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104157 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104177 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104195 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104215 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104255 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104271 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104288 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104308 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104335 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104402 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104414 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104426 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104435 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104444 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104453 4754 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104461 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104471 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104480 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104489 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104498 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104507 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104516 4754 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104525 4754 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104534 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104531 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104543 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104616 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104637 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104655 4754 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104671 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104683 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104694 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104704 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104716 4754 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104729 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104742 4754 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104756 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104768 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104781 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104790 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104802 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104824 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104840 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104853 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104868 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104883 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104895 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104908 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104921 4754 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104936 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104949 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104985 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104998 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105012 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105026 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105043 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105055 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105069 4754 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105082 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105095 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105109 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105121 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105134 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105147 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105159 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105171 4754 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105183 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105197 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105209 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105222 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105236 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105248 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105257 4754 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105268 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105278 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105287 4754 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105296 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105305 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105314 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105323 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105333 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105342 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105351 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105361 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105371 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105381 4754 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105390 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105400 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105411 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105420 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105429 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105438 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105447 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105456 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105468 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104842 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104929 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105488 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.104913 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105048 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105347 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105358 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105373 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105545 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105816 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105866 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105871 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.105954 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106200 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106248 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106282 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106379 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106529 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106700 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.106805 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.107266 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.107339 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:12.60731818 +0000 UTC m=+20.166262965 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107342 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107519 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107624 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107708 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.107773 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.107938 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.108034 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:12.608016048 +0000 UTC m=+20.166960943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.108766 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109007 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109039 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109291 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109510 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109712 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109869 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109877 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.110247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.110407 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.110648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.110830 4754 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.110896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.111311 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.111715 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112034 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.109488 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112441 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112478 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112653 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112840 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.112659 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.113281 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.113838 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.114913 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.114996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.115206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.116255 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.116526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.117006 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.116862 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.117448 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.118064 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.118759 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.120596 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.120603 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.121319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.121470 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.121674 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.122872 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.124108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.124283 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.125196 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.126552 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.127107 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.127453 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127620 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127641 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127652 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127731 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:12.627715913 +0000 UTC m=+20.186660688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127879 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.127980 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.128072 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.128245 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:12.628166715 +0000 UTC m=+20.187111500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131285 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131321 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131447 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131541 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131598 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131839 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131901 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131986 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.131666 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.132682 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.132746 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.132785 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.133179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.133378 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.133500 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134158 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134614 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134773 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.134892 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.135212 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.135307 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.135727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.136058 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.136116 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.136062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.136164 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.137067 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.137735 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.138325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.138397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.142023 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.142129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.145734 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.145916 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.152374 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.166217 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.171496 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.176799 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.179618 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4" exitCode=255 Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.179686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4"} Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.188426 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.191744 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.191829 4754 scope.go:117] "RemoveContainer" containerID="3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.198459 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.207849 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.208355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.208554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.208824 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209032 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209048 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209059 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209070 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209082 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209094 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209107 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209117 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209068 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209128 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209181 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209212 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209222 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209234 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209245 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209256 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209267 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209276 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209284 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209292 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209303 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209312 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209323 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209336 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209347 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209361 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209370 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209379 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209387 4754 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209396 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209406 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209418 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209430 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209443 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209458 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209469 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209478 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209486 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209499 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.209651 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210199 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210213 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210246 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210258 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210266 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210276 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210285 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210293 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210300 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210321 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210330 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210338 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210347 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210355 4754 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210363 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210372 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210381 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210390 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210399 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210410 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210420 4754 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210429 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210437 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210445 4754 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210457 4754 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210466 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210476 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210505 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210514 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210523 4754 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210532 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210600 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210612 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210622 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210632 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210640 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210651 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210662 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210674 4754 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210736 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210745 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210756 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210800 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210815 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210827 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210835 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210921 4754 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210956 4754 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.210999 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.214892 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.214908 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215392 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215450 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215461 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215470 4754 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215479 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215488 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215497 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215506 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215532 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215543 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215566 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215576 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215600 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215611 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215621 4754 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215629 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215643 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215652 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.215661 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.216903 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.226020 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.235541 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.334331 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.344145 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.349612 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:06:12 crc kubenswrapper[4754]: W1011 03:06:12.356680 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8cd7da3d28c7ced99bcc13659c7188bbf1f161fa9c32fe24d39348057e9be11f WatchSource:0}: Error finding container 8cd7da3d28c7ced99bcc13659c7188bbf1f161fa9c32fe24d39348057e9be11f: Status 404 returned error can't find the container with id 8cd7da3d28c7ced99bcc13659c7188bbf1f161fa9c32fe24d39348057e9be11f Oct 11 03:06:12 crc kubenswrapper[4754]: W1011 03:06:12.367461 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7c93114efee7b74cc12a4189957abfe4a50e927c0f42a7fe6481d5b21dd3d3ad WatchSource:0}: Error finding container 7c93114efee7b74cc12a4189957abfe4a50e927c0f42a7fe6481d5b21dd3d3ad: Status 404 returned error can't find the container with id 7c93114efee7b74cc12a4189957abfe4a50e927c0f42a7fe6481d5b21dd3d3ad Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.618731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.618818 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.618847 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.618911 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:13.618894566 +0000 UTC m=+21.177839351 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.618933 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.619022 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.619062 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:13.619038209 +0000 UTC m=+21.177982994 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.619078 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:13.61907109 +0000 UTC m=+21.178015865 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.719640 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:12 crc kubenswrapper[4754]: I1011 03:06:12.719676 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719828 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719865 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719837 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719878 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719886 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719897 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.719924 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:13.719909307 +0000 UTC m=+21.278854092 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:12 crc kubenswrapper[4754]: E1011 03:06:12.720000 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:13.719941918 +0000 UTC m=+21.278886723 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.081701 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.088734 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.089934 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.093030 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.095210 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.098045 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.099177 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.101088 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.103235 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.104751 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.107059 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.108199 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.110378 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.110911 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.112106 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.112676 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.112692 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.113989 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.114762 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.115248 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.116390 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.117085 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.117639 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.118693 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.119159 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.120271 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.120673 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.121748 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.122441 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.124258 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.124806 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.125769 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.126240 4754 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.126343 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.128228 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.129081 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.129531 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.131172 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.132208 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.132692 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.133689 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.134132 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.134442 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.135404 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.136121 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.137104 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.137758 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.138823 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.139357 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.140222 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.140940 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.142213 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.142661 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.143519 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.144201 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.144823 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.146419 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.147202 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.147240 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.153994 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.169283 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.187184 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.187270 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.187294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7c93114efee7b74cc12a4189957abfe4a50e927c0f42a7fe6481d5b21dd3d3ad"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.188951 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8cd7da3d28c7ced99bcc13659c7188bbf1f161fa9c32fe24d39348057e9be11f"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.191111 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.191205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5e77c91840f97e145b0376dabd64ffe697af18b365f60314df9807dbb37cdaed"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.193804 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.194442 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.196732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270"} Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.197534 4754 status_manager.go:317] "Container readiness changed for unknown container" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.197636 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.211409 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.226338 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.240263 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.256610 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.275795 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.294562 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.308713 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.325481 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.346260 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.367779 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.382017 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.398022 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.414816 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.449232 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.472127 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.486178 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.502320 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.519393 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.633694 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.633794 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.633848 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.633891 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:15.633860853 +0000 UTC m=+23.192805638 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.633951 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.634026 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.634032 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:15.634013647 +0000 UTC m=+23.192958492 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.634080 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:15.634072008 +0000 UTC m=+23.193016793 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.734876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:13 crc kubenswrapper[4754]: I1011 03:06:13.735243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735021 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735362 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735376 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735424 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:15.735408988 +0000 UTC m=+23.294353783 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735344 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735675 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735731 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:13 crc kubenswrapper[4754]: E1011 03:06:13.735810 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:15.735800948 +0000 UTC m=+23.294745733 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.083521 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.083588 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:14 crc kubenswrapper[4754]: E1011 03:06:14.083665 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:14 crc kubenswrapper[4754]: E1011 03:06:14.083778 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.084005 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:14 crc kubenswrapper[4754]: E1011 03:06:14.084214 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.199223 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.538110 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.552446 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.560328 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.563491 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.570494 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.590762 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.609131 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.624089 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.637748 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.657476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.671944 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.686991 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.698881 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.719438 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.734792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.753990 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.770041 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.783373 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.796427 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:14 crc kubenswrapper[4754]: I1011 03:06:14.811721 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:14Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:15 crc kubenswrapper[4754]: I1011 03:06:15.652195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:15 crc kubenswrapper[4754]: I1011 03:06:15.652366 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:15 crc kubenswrapper[4754]: I1011 03:06:15.652419 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.652601 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.652696 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:19.652669155 +0000 UTC m=+27.211613970 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.653243 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.653403 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:19.653360653 +0000 UTC m=+27.212305458 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.653551 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:19.653494376 +0000 UTC m=+27.212439201 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:15 crc kubenswrapper[4754]: I1011 03:06:15.753370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:15 crc kubenswrapper[4754]: I1011 03:06:15.753444 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753687 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753699 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753766 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753790 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753719 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753906 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.753882 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:19.75384785 +0000 UTC m=+27.312792665 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:15 crc kubenswrapper[4754]: E1011 03:06:15.754101 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:19.754036835 +0000 UTC m=+27.312981660 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.083655 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.083779 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.083663 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:16 crc kubenswrapper[4754]: E1011 03:06:16.083881 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:16 crc kubenswrapper[4754]: E1011 03:06:16.084095 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:16 crc kubenswrapper[4754]: E1011 03:06:16.084315 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.209222 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3"} Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.232324 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.254837 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.281187 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.303160 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.326571 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.356194 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.375639 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.393331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:16 crc kubenswrapper[4754]: I1011 03:06:16.408505 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:16Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.015895 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-pmk28"] Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.016206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.018074 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.018690 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.022539 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.048356 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.066723 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.070358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-hosts-file\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.070506 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xns6\" (UniqueName: \"kubernetes.io/projected/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-kube-api-access-9xns6\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.083285 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.083375 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.083465 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.083544 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.083705 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.083954 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.086239 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.100008 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.115423 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.127392 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.149605 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.163365 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.171315 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-hosts-file\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.171359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xns6\" (UniqueName: \"kubernetes.io/projected/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-kube-api-access-9xns6\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.171641 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-hosts-file\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.178309 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.192108 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.195043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xns6\" (UniqueName: \"kubernetes.io/projected/ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68-kube-api-access-9xns6\") pod \"node-resolver-pmk28\" (UID: \"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\") " pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.289134 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.292139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.292240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.292270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.292408 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.301919 4754 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.302395 4754 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.304060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.304182 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.304279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.304392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.304517 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.327175 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pmk28" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.331420 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.337172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.337294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.337311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.337332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.337345 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: W1011 03:06:18.340795 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecb2bdbe_2019_41c4_ada8_c5aa0d32dd68.slice/crio-e457edede0e34384b6a4f3355be72a24c3b9038590a7371bb65cf3be6a95b5f1 WatchSource:0}: Error finding container e457edede0e34384b6a4f3355be72a24c3b9038590a7371bb65cf3be6a95b5f1: Status 404 returned error can't find the container with id e457edede0e34384b6a4f3355be72a24c3b9038590a7371bb65cf3be6a95b5f1 Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.359486 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.370696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.370767 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.370787 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.370818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.370840 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.391234 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.398075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.398131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.398152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.398184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.398205 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.417838 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.428370 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-42sbf"] Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.429847 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.442463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.442507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.442520 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.442541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.442556 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.443163 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.443452 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2nlvv"] Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.443582 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.443808 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-px2sf"] Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444142 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-r59xj"] Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444180 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.443821 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444047 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444183 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444263 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444874 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.444916 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.445086 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.463229 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.463629 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.464082 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.464324 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.464502 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.464704 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.464963 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.465362 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.465685 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.466015 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.466591 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.466492 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.466816 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.470764 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: E1011 03:06:18.470880 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-conf-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-system-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474159 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cni-binary-copy\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-mcd-auth-proxy-config\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474223 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474249 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-socket-dir-parent\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474316 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cnibin\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474346 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txjv7\" (UniqueName: \"kubernetes.io/projected/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-kube-api-access-txjv7\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474400 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474425 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-bin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474456 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-binary-copy\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474510 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9bs6\" (UniqueName: \"kubernetes.io/projected/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-kube-api-access-b9bs6\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-netns\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-etc-kubernetes\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474593 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474621 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474657 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474720 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cnibin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474808 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-hostroot\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474903 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.474941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.475005 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-kubelet\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.476817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.476868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.476883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.476906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.476921 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.478840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f5mb\" (UniqueName: \"kubernetes.io/projected/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-kube-api-access-6f5mb\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.479956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-system-cni-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480092 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-os-release\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-daemon-config\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480161 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-multus-certs\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480192 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480223 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-proxy-tls\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480245 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480282 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480335 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4t5f\" (UniqueName: \"kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480361 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-os-release\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480414 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480439 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-k8s-cni-cncf-io\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480465 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480491 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-rootfs\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480547 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-multus\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.480627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.496374 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.515337 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.533753 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.549572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.562614 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.575484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.580515 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.580571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.580586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.580610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.580625 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581104 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-kubelet\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581140 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f5mb\" (UniqueName: \"kubernetes.io/projected/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-kube-api-access-6f5mb\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581185 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-system-cni-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581214 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581246 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-os-release\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581274 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-daemon-config\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-multus-certs\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-proxy-tls\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581387 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581430 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581474 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4t5f\" (UniqueName: \"kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581506 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-os-release\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581663 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-k8s-cni-cncf-io\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581720 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581749 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-rootfs\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-multus\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-conf-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581904 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-system-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.581944 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cni-binary-copy\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-mcd-auth-proxy-config\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-socket-dir-parent\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582129 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582181 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cnibin\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582276 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txjv7\" (UniqueName: \"kubernetes.io/projected/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-kube-api-access-txjv7\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582318 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582393 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-bin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582426 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582465 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-binary-copy\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582504 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9bs6\" (UniqueName: \"kubernetes.io/projected/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-kube-api-access-b9bs6\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-netns\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582581 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-etc-kubernetes\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582619 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582656 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582690 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582725 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cnibin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-hostroot\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582826 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.582904 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583051 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583122 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583216 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-os-release\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.583744 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584144 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-system-cni-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584198 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584235 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-k8s-cni-cncf-io\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584264 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584314 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-rootfs\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584337 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-multus\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584360 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-kubelet\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-conf-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584541 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-system-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584540 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-socket-dir-parent\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-multus-certs\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.584934 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585070 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585188 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cni-binary-copy\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-var-lib-cni-bin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-cni-dir\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585665 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585629 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585655 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-hostroot\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585680 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cnibin\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585779 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585865 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-host-run-netns\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.585893 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-etc-kubernetes\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586063 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586646 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-os-release\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-cnibin\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.586792 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.587368 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-multus-daemon-config\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.587499 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-cni-binary-copy\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.587589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.589056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-mcd-auth-proxy-config\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.590019 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-proxy-tls\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.591587 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.601604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4t5f\" (UniqueName: \"kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f\") pod \"ovnkube-node-42sbf\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.601941 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txjv7\" (UniqueName: \"kubernetes.io/projected/ef866bf9-b2ed-4ab9-922b-68f7373b6c04-kube-api-access-txjv7\") pod \"machine-config-daemon-px2sf\" (UID: \"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\") " pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.605206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9bs6\" (UniqueName: \"kubernetes.io/projected/ef19b293-a4d7-44f5-a26d-4daf0e558d6d-kube-api-access-b9bs6\") pod \"multus-additional-cni-plugins-r59xj\" (UID: \"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\") " pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.605633 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f5mb\" (UniqueName: \"kubernetes.io/projected/3cd26d5b-a1b6-41fa-b285-d650e389d8d4-kube-api-access-6f5mb\") pod \"multus-2nlvv\" (UID: \"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\") " pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.611501 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.637894 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.651631 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.665909 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683587 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.683882 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.700980 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.718302 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.732686 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.754154 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.756172 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.770189 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: W1011 03:06:18.771925 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dea8260_bc1e_4ad6_96d3_f9d202b125c7.slice/crio-3fe3d31161b7a0f6a9dfb1bf95acfe17ad3d904f7941e3cb2aa64a7294ce0c80 WatchSource:0}: Error finding container 3fe3d31161b7a0f6a9dfb1bf95acfe17ad3d904f7941e3cb2aa64a7294ce0c80: Status 404 returned error can't find the container with id 3fe3d31161b7a0f6a9dfb1bf95acfe17ad3d904f7941e3cb2aa64a7294ce0c80 Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.780499 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2nlvv" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.787080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.787121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.787134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.787152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.787165 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.791229 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-r59xj" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.804286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:06:18 crc kubenswrapper[4754]: W1011 03:06:18.806476 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cd26d5b_a1b6_41fa_b285_d650e389d8d4.slice/crio-6916dc8f26ac52287f709da1ae80df227c7ec433125ccb31ad3486bf49bfc91d WatchSource:0}: Error finding container 6916dc8f26ac52287f709da1ae80df227c7ec433125ccb31ad3486bf49bfc91d: Status 404 returned error can't find the container with id 6916dc8f26ac52287f709da1ae80df227c7ec433125ccb31ad3486bf49bfc91d Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.806911 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.828699 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.843914 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.874384 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.902376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.902424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.902463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.902486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.902503 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:18Z","lastTransitionTime":"2025-10-11T03:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.928479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.948724 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:18 crc kubenswrapper[4754]: I1011 03:06:18.958492 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:18Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.005740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.005770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.005780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.005798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.005807 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.109239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.109686 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.109696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.109717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.109731 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.212388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.212438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.212452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.212481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.212498 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.219565 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" exitCode=0 Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.219642 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.219693 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"3fe3d31161b7a0f6a9dfb1bf95acfe17ad3d904f7941e3cb2aa64a7294ce0c80"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.221340 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pmk28" event={"ID":"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68","Type":"ContainerStarted","Data":"0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.221408 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pmk28" event={"ID":"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68","Type":"ContainerStarted","Data":"e457edede0e34384b6a4f3355be72a24c3b9038590a7371bb65cf3be6a95b5f1"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.223084 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerStarted","Data":"acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.223115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerStarted","Data":"6916dc8f26ac52287f709da1ae80df227c7ec433125ccb31ad3486bf49bfc91d"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.229465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.229499 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.229513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"644a5b2c90b5b5c44f7666e6d3dd05c439b52ca36775fddf6ca361738b45b918"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.236493 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerStarted","Data":"b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.236566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerStarted","Data":"a679c3755cb15d893e2070d1bff227a128bec09af0cbd47dad03a5427c159a5d"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.240059 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.260820 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.273819 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.296249 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.314453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.315009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.315101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.315169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.315245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.315300 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.334417 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.351337 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.368804 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.382389 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.405893 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.419101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.419137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.419145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.419160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.419169 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.422668 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.438327 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.451479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.463160 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.475904 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.496545 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.513680 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.521660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.521810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.521892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.521991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.522089 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.538535 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.551674 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.574333 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.592342 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.614196 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.625141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.625175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.625187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.625204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.625216 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.631954 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.658749 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.672655 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.694797 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.699256 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.699402 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:27.699374677 +0000 UTC m=+35.258319612 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.699564 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.699687 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.699745 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.699826 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.699834 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:27.699807949 +0000 UTC m=+35.258752914 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.699892 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:27.699881311 +0000 UTC m=+35.258826096 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.714930 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.727308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.727345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.727358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.727380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.727394 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.739678 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.806987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.807061 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807235 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807259 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807276 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807351 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:27.80733186 +0000 UTC m=+35.366276645 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807612 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807731 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807823 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:19 crc kubenswrapper[4754]: E1011 03:06:19.807996 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:27.807978077 +0000 UTC m=+35.366922862 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.831454 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.831712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.831783 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.831853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.831919 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.935404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.935993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.936012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.936036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:19 crc kubenswrapper[4754]: I1011 03:06:19.936054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:19Z","lastTransitionTime":"2025-10-11T03:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.041353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.041403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.041433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.041461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.041479 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.083260 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.083418 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.083605 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:20 crc kubenswrapper[4754]: E1011 03:06:20.083600 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:20 crc kubenswrapper[4754]: E1011 03:06:20.083791 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:20 crc kubenswrapper[4754]: E1011 03:06:20.083912 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.144158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.144208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.144220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.144237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.144247 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.242902 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904" exitCode=0 Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.243040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.245787 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.245809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.245817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.245830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.245842 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248306 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248429 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248508 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248591 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248668 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.248794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.262059 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.283319 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.296544 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.313936 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.337414 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.349203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.349236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.349245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.349262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.349273 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.351998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.367836 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.385285 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.412484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.428980 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.444602 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.453011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.453040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.453053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.453365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.455319 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.459639 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.473420 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.486286 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:20Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.558185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.558226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.558242 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.558259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.558272 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.661306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.661558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.661581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.661610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.661638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.765176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.765238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.765253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.765275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.765290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.868612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.868665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.868679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.868706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.868721 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.972042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.972086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.972096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.972112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:20 crc kubenswrapper[4754]: I1011 03:06:20.972123 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:20Z","lastTransitionTime":"2025-10-11T03:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.076429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.076912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.076945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.077012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.077033 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.180639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.182179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.182222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.182260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.182284 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.257607 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6" exitCode=0 Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.257673 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.286394 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.286464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.286495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.286535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.286563 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.288844 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.307641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.336196 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.360010 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.389326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.389412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.389434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.389468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.389494 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.396996 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.414744 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.431738 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.446841 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.467865 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.482572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.492529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.492566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.492580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.492601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.492617 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.499561 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.521334 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.538453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.551351 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.595755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.595802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.595814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.595829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.595841 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.698694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.698737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.698747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.698763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.698775 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.802303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.802351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.802362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.802380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.802398 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.904817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.904864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.904875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.904892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:21 crc kubenswrapper[4754]: I1011 03:06:21.904904 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:21Z","lastTransitionTime":"2025-10-11T03:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.007468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.007509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.007528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.007544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.007554 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.083005 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.083053 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.083012 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:22 crc kubenswrapper[4754]: E1011 03:06:22.083129 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:22 crc kubenswrapper[4754]: E1011 03:06:22.083214 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:22 crc kubenswrapper[4754]: E1011 03:06:22.083299 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.110377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.110420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.110429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.110444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.110457 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.213294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.213333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.213341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.213359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.213370 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.261848 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c" exitCode=0 Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.261912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.270143 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.286392 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.303040 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.315677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.315740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.315753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.315776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.315788 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.318746 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.329813 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.348037 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.364902 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.384362 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.406532 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.447351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.447457 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.447507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.447529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.447545 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.451641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.464911 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.480634 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.493814 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.506877 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.521070 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.551117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.551161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.551170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.551188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.551197 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.654768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.655369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.655428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.655456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.655510 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.758677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.758766 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.758795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.758830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.758853 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.862619 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.862668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.862682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.862698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.862711 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.965947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.966004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.966013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.966030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:22 crc kubenswrapper[4754]: I1011 03:06:22.966039 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:22Z","lastTransitionTime":"2025-10-11T03:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.031100 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-nxj2n"] Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.031850 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.035207 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.035377 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.035867 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.037498 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.050327 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.064015 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.068552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.068596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.068609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.068627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.068639 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.079608 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.096331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.117791 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.130911 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.145288 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.153503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eb406ab-5a49-485b-a5a5-af1d48675ae8-host\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.153544 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4eb406ab-5a49-485b-a5a5-af1d48675ae8-serviceca\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.153587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff75d\" (UniqueName: \"kubernetes.io/projected/4eb406ab-5a49-485b-a5a5-af1d48675ae8-kube-api-access-ff75d\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.158428 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177218 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.177513 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.199464 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.218158 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.240898 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.254279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.254486 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eb406ab-5a49-485b-a5a5-af1d48675ae8-host\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.254533 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4eb406ab-5a49-485b-a5a5-af1d48675ae8-serviceca\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.254588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff75d\" (UniqueName: \"kubernetes.io/projected/4eb406ab-5a49-485b-a5a5-af1d48675ae8-kube-api-access-ff75d\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.254617 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eb406ab-5a49-485b-a5a5-af1d48675ae8-host\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.256171 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4eb406ab-5a49-485b-a5a5-af1d48675ae8-serviceca\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.267974 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.275060 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff75d\" (UniqueName: \"kubernetes.io/projected/4eb406ab-5a49-485b-a5a5-af1d48675ae8-kube-api-access-ff75d\") pod \"node-ca-nxj2n\" (UID: \"4eb406ab-5a49-485b-a5a5-af1d48675ae8\") " pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.280074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.280186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.280215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.280241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.280259 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.282405 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b" exitCode=0 Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.282448 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.285045 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.299664 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.311407 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.322819 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.340048 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.348475 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nxj2n" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.355670 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.367282 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: W1011 03:06:23.373496 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eb406ab_5a49_485b_a5a5_af1d48675ae8.slice/crio-7b63e525d1ff6e9f7240fb88cfd4f01ded3c9a81174e652d567ebeb6435fb44d WatchSource:0}: Error finding container 7b63e525d1ff6e9f7240fb88cfd4f01ded3c9a81174e652d567ebeb6435fb44d: Status 404 returned error can't find the container with id 7b63e525d1ff6e9f7240fb88cfd4f01ded3c9a81174e652d567ebeb6435fb44d Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.381284 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.383253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.383333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.383349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.383372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.383433 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.398768 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.419635 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.435823 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.450284 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.474814 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.486265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.486312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.486323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.486343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.486359 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.492218 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.517327 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.535606 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.549518 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.565102 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.578653 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.588092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.588127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.588136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.588153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.588163 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.596860 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.613015 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.644016 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.658356 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.671581 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.682997 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.690463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.690500 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.690510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.690528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.690541 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.699832 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.712844 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.723522 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.736408 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.748354 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.760889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.792722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.792767 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.792778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.792796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.792809 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.895216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.895252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.895260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.895275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.895287 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.998200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.998244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.998255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.998271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:23 crc kubenswrapper[4754]: I1011 03:06:23.998282 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:23Z","lastTransitionTime":"2025-10-11T03:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.082896 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.082922 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:24 crc kubenswrapper[4754]: E1011 03:06:24.083071 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:24 crc kubenswrapper[4754]: E1011 03:06:24.083214 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.083422 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:24 crc kubenswrapper[4754]: E1011 03:06:24.083499 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.100516 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.100986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.101002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.101020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.101033 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.204245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.204321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.204340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.204370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.204389 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.288168 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nxj2n" event={"ID":"4eb406ab-5a49-485b-a5a5-af1d48675ae8","Type":"ContainerStarted","Data":"49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.288291 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nxj2n" event={"ID":"4eb406ab-5a49-485b-a5a5-af1d48675ae8","Type":"ContainerStarted","Data":"7b63e525d1ff6e9f7240fb88cfd4f01ded3c9a81174e652d567ebeb6435fb44d"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.300230 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55" exitCode=0 Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.300318 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.305214 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.307420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.307466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.307489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.307526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.307549 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.319331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.342581 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.362019 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.383624 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.406800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.417156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.417222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.417240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.417264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.417280 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.424271 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.448280 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.465273 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.480542 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.494014 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.505889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521357 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.521737 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.534255 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.546743 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.565289 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.580863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.593822 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.605629 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.622253 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.624179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.624213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.624226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.624247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.624263 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.633608 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.646901 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.659518 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.672749 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.686372 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.698496 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.717334 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.727447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.727494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.727506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.727525 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.727540 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.748462 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.772340 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.794104 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:24Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.830623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.830672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.830683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.830705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.830719 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.934391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.934465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.934483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.934517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:24 crc kubenswrapper[4754]: I1011 03:06:24.934543 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:24Z","lastTransitionTime":"2025-10-11T03:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.038711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.038783 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.038801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.038831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.038850 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.141837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.141908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.141932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.142011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.142038 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.245012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.245089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.245113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.245144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.245168 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.318952 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.319626 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.325673 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef19b293-a4d7-44f5-a26d-4daf0e558d6d" containerID="346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8" exitCode=0 Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.325748 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerDied","Data":"346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.341384 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.348503 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.348572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.348595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.348627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.348648 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.367960 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.373991 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.387755 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.407348 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.439233 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.452470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.452521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.452534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.452557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.452571 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.461215 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.478056 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.493949 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.515670 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.533875 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.548595 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.555016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.555053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.555061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.555076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.555086 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.561999 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.575188 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.586238 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.596667 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.609488 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.620357 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.632496 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.644282 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.658316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.658358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.658370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.658389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.658401 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.662990 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.674638 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.685039 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.694751 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.710673 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.721920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.733143 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.744782 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.755732 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.761092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.761144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.761163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.761189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.761207 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.766401 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.775722 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:25Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.869725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.869787 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.869798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.869815 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.869825 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.971912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.972346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.972366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.972385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:25 crc kubenswrapper[4754]: I1011 03:06:25.972398 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:25Z","lastTransitionTime":"2025-10-11T03:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.073986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.074060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.074069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.074083 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.074092 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.083427 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.083498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.083440 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:26 crc kubenswrapper[4754]: E1011 03:06:26.083672 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:26 crc kubenswrapper[4754]: E1011 03:06:26.083865 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:26 crc kubenswrapper[4754]: E1011 03:06:26.084024 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.177172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.177305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.177324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.177350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.177370 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.280254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.280545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.280627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.280784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.280868 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.337009 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" event={"ID":"ef19b293-a4d7-44f5-a26d-4daf0e558d6d","Type":"ContainerStarted","Data":"8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.338272 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.338299 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.375691 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.376945 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.383878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.384393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.384611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.384777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.385013 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.399196 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.414485 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.429902 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.451231 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.468742 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.487289 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.489603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.489659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.489679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.489703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.489720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.508119 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.531294 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.552223 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.569636 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.587103 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.592210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.592260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.592279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.593181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.593202 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.605633 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.629581 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.655196 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.668878 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.687213 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.695862 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.696228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.696259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.696280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.696295 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.705212 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.721927 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.747307 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.762891 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.782889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.799501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.799559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.799576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.799600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.799616 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.800376 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.823754 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.839995 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.862594 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.881584 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.898808 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.903081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.903139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.903157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.903180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.903194 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:26Z","lastTransitionTime":"2025-10-11T03:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.914166 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:26 crc kubenswrapper[4754]: I1011 03:06:26.926818 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:26Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.005592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.005665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.005682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.005710 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.005728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.108498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.108578 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.108603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.108631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.108656 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.211688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.211758 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.211776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.211803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.211822 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.314690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.314760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.314776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.314792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.314804 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.341998 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/0.log" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.345238 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece" exitCode=1 Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.345280 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.345881 4754 scope.go:117] "RemoveContainer" containerID="4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.357226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.369131 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.380036 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.394990 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.416868 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.418261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.418335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.418345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.418381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.418392 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.429107 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.442193 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.457053 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.474933 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"message\\\":\\\"kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:27.281326 6040 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281406 6040 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281524 6040 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1011 03:06:27.282109 6040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:27.282174 6040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:27.282191 6040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1011 03:06:27.282196 6040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1011 03:06:27.282222 6040 factory.go:656] Stopping watch factory\\\\nI1011 03:06:27.282217 6040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:27.282232 6040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.488339 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.508279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.521049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.521101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.521115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.521134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.521148 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.529846 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.549936 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.564342 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.577454 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.623645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.623681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.623690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.623705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.623716 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.707439 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.707631 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:06:43.707603205 +0000 UTC m=+51.266548020 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.707710 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.707770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.707836 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.707860 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.707912 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:43.707899893 +0000 UTC m=+51.266844698 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.707938 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:43.707928554 +0000 UTC m=+51.266873359 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.731381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.731435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.731446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.731463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.731474 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.781397 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.795071 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.808267 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.808612 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.808644 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.808754 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.808777 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.808788 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.808831 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:43.808818842 +0000 UTC m=+51.367763627 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.809152 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.809169 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.809177 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:27 crc kubenswrapper[4754]: E1011 03:06:27.809198 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:43.809191651 +0000 UTC m=+51.368136426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.824327 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.833302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.833327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.833335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.833349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.833359 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.839671 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.852148 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.865294 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.879467 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.897985 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"message\\\":\\\"kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:27.281326 6040 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281406 6040 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281524 6040 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1011 03:06:27.282109 6040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:27.282174 6040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:27.282191 6040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1011 03:06:27.282196 6040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1011 03:06:27.282222 6040 factory.go:656] Stopping watch factory\\\\nI1011 03:06:27.282217 6040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:27.282232 6040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.914357 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.935538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.935589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.935599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.935615 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.935624 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:27Z","lastTransitionTime":"2025-10-11T03:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.938320 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.951283 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.964141 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.975847 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:27 crc kubenswrapper[4754]: I1011 03:06:27.986807 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:27Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.002181 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.039613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.039653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.039665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.039683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.039696 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.083242 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.083303 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.083261 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.083369 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.083426 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.083496 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.142075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.142123 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.142135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.142151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.142163 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.245044 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.245111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.245122 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.245149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.245160 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.347755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.347794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.347806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.347820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.347831 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.350272 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/1.log" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.351168 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/0.log" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.353837 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad" exitCode=1 Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.353879 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.353920 4754 scope.go:117] "RemoveContainer" containerID="4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.355167 4754 scope.go:117] "RemoveContainer" containerID="ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad" Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.355504 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.373318 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.386485 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.404464 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.431848 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.448481 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.449629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.449674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.449691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.449709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.449723 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.460411 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.470998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.487922 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4458a4333bc8704b970a4366f4f67401d0b69a201ea74dd5b050a52cdb5a0ece\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"message\\\":\\\"kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:27.281326 6040 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281406 6040 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:27.281524 6040 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1011 03:06:27.282109 6040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:27.282174 6040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:27.282191 6040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1011 03:06:27.282196 6040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1011 03:06:27.282222 6040 factory.go:656] Stopping watch factory\\\\nI1011 03:06:27.282217 6040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:27.282232 6040 handler.go:208] Removed *v1.Node event handler 2\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1011 03:06:27.282248 6040 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.504414 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.517733 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.531280 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.543107 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.552342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.552382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.552394 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.552414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.552428 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.553415 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.562539 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.573023 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.588329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.588367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.588383 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.588404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.588419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.601110 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.604571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.604613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.604628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.604651 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.604665 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.617532 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.621222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.621258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.621269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.621283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.621294 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.634128 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.637369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.637401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.637412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.637428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.637441 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.648026 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.651329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.651353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.651363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.651379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.651413 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.662419 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:28Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:28 crc kubenswrapper[4754]: E1011 03:06:28.662536 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.664000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.664026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.664034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.664048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.664058 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.766948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.766998 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.767009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.767027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.767038 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.869124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.869165 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.869177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.869192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.869204 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.971498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.971559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.971572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.971587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:28 crc kubenswrapper[4754]: I1011 03:06:28.971596 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:28Z","lastTransitionTime":"2025-10-11T03:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.074672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.074733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.074750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.074777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.074794 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.176636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.176707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.176723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.176744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.176757 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.278478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.278521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.278530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.278546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.278555 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.359751 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/1.log" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.363371 4754 scope.go:117] "RemoveContainer" containerID="ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad" Oct 11 03:06:29 crc kubenswrapper[4754]: E1011 03:06:29.363584 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.379161 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.381110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.381138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.381149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.381166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.381178 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.392773 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.404532 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.418701 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.431878 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.441118 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.454947 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.469811 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.481755 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.483662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.483712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.483722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.483737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.483747 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.495425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.512638 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.525332 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.545114 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.557830 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.571149 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.585917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.586027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.586039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.586056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.586067 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.688903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.688947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.688958 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.688994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.689007 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.791711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.791767 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.791809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.791831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.791845 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.894139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.894187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.894201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.894218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.894233 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.996531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.996582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.996596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.996616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:29 crc kubenswrapper[4754]: I1011 03:06:29.996629 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:29Z","lastTransitionTime":"2025-10-11T03:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.082646 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.082697 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.082703 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:30 crc kubenswrapper[4754]: E1011 03:06:30.082799 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:30 crc kubenswrapper[4754]: E1011 03:06:30.082918 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:30 crc kubenswrapper[4754]: E1011 03:06:30.083099 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.099637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.099690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.099720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.099739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.099751 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.202947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.203038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.203052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.203076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.203091 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.306037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.306072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.306081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.306096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.306106 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.409089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.409179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.409208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.409243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.409266 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.512667 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.512751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.512771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.512812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.512832 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.615312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.615378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.615396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.615424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.615443 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.719082 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.719128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.719141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.719159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.719172 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.821941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.822051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.822077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.822114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.822138 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.823612 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p"] Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.824374 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:30 crc kubenswrapper[4754]: W1011 03:06:30.826198 4754 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 11 03:06:30 crc kubenswrapper[4754]: E1011 03:06:30.826270 4754 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 03:06:30 crc kubenswrapper[4754]: W1011 03:06:30.827399 4754 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 11 03:06:30 crc kubenswrapper[4754]: E1011 03:06:30.827469 4754 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.843029 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.859486 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.874994 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.890733 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.907179 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.924886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.924942 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.924957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.925001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.925015 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:30Z","lastTransitionTime":"2025-10-11T03:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.928204 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.945957 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.948540 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.948685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93810265-58d5-4b44-a368-8b78981b9f2f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.948842 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.948903 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkrbv\" (UniqueName: \"kubernetes.io/projected/93810265-58d5-4b44-a368-8b78981b9f2f-kube-api-access-xkrbv\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.959776 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.977676 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:30 crc kubenswrapper[4754]: I1011 03:06:30.996143 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:30Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.012879 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.027837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.027891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.027905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.027934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.027950 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.029174 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.047038 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.050556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.050618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkrbv\" (UniqueName: \"kubernetes.io/projected/93810265-58d5-4b44-a368-8b78981b9f2f-kube-api-access-xkrbv\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.050682 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.050728 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93810265-58d5-4b44-a368-8b78981b9f2f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.051897 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.052279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93810265-58d5-4b44-a368-8b78981b9f2f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.067300 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.072662 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkrbv\" (UniqueName: \"kubernetes.io/projected/93810265-58d5-4b44-a368-8b78981b9f2f-kube-api-access-xkrbv\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.088395 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.107471 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.131358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.131413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.131436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.131470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.131493 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.234860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.234918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.234930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.234953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.234982 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.337697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.337777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.337799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.337825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.337843 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.441462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.441540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.441552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.441593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.441609 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.544837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.544887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.544901 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.544922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.544935 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.647804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.647882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.647899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.647927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.647946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.751388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.751461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.751479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.751513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.751543 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.851376 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.855014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.855063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.855147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.855702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.855745 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.951546 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b6s82"] Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.952404 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:31 crc kubenswrapper[4754]: E1011 03:06:31.952508 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.953051 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.969062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.969605 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.969643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.969740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.969824 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:31Z","lastTransitionTime":"2025-10-11T03:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.974410 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.977043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93810265-58d5-4b44-a368-8b78981b9f2f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nrg7p\" (UID: \"93810265-58d5-4b44-a368-8b78981b9f2f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:31 crc kubenswrapper[4754]: I1011 03:06:31.998419 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.016916 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.039026 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.042621 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.060119 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: W1011 03:06:32.063150 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93810265_58d5_4b44_a368_8b78981b9f2f.slice/crio-1940555e7169a6c8023a546bc11ab14d005c3ff21dbe316e57e55bfbf9231bdc WatchSource:0}: Error finding container 1940555e7169a6c8023a546bc11ab14d005c3ff21dbe316e57e55bfbf9231bdc: Status 404 returned error can't find the container with id 1940555e7169a6c8023a546bc11ab14d005c3ff21dbe316e57e55bfbf9231bdc Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.067835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.067873 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tbx5\" (UniqueName: \"kubernetes.io/projected/fb5e7151-ac6f-4da4-9cc8-39639136e005-kube-api-access-6tbx5\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.074134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.074196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.074212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.074233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.074246 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.083379 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.083415 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.083555 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.083531 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.083620 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.083712 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.083774 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.098353 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.117273 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.132271 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.147685 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.160292 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.168710 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.168788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tbx5\" (UniqueName: \"kubernetes.io/projected/fb5e7151-ac6f-4da4-9cc8-39639136e005-kube-api-access-6tbx5\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.169076 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.170465 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:32.670425108 +0000 UTC m=+40.229369943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.172693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.177705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.177762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.177783 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.177812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.177834 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.186053 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.194362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tbx5\" (UniqueName: \"kubernetes.io/projected/fb5e7151-ac6f-4da4-9cc8-39639136e005-kube-api-access-6tbx5\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.201027 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.214147 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.257688 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.272588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.281360 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.281409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.281422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.281445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.281459 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.373594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" event={"ID":"93810265-58d5-4b44-a368-8b78981b9f2f","Type":"ContainerStarted","Data":"f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.373644 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" event={"ID":"93810265-58d5-4b44-a368-8b78981b9f2f","Type":"ContainerStarted","Data":"1940555e7169a6c8023a546bc11ab14d005c3ff21dbe316e57e55bfbf9231bdc"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.384343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.384381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.384393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.384408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.384422 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.487206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.487278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.487297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.487327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.487346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.590739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.590780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.590793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.590813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.590823 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.673920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.674139 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:32 crc kubenswrapper[4754]: E1011 03:06:32.674252 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:33.67423152 +0000 UTC m=+41.233176305 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.694724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.694840 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.694870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.694909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.694932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.798217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.798272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.798287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.798307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.798322 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.902635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.903075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.903094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.903121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:32 crc kubenswrapper[4754]: I1011 03:06:32.903143 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:32Z","lastTransitionTime":"2025-10-11T03:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.006743 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.006825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.006844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.006877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.006898 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.083328 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:33 crc kubenswrapper[4754]: E1011 03:06:33.083610 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.107067 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.116121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.116257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.116290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.116344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.116369 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.135483 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.163481 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.181496 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.199424 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.220388 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.222092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.222151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.222166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.222189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.222212 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.242255 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.266365 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.290369 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.318115 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.324869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.324925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.324938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.324985 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.325004 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.337220 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.370957 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.379821 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" event={"ID":"93810265-58d5-4b44-a368-8b78981b9f2f","Type":"ContainerStarted","Data":"ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.397031 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.418181 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.429029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.429129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.429157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.429201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.429228 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.440172 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.460531 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.478312 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.493288 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.508941 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.524274 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.533127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.533182 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.533193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.533210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.533221 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.538787 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.552862 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.569269 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.582416 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.602922 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.620941 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.635583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.635637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.635657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.635688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.635708 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.643533 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.663363 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.682328 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.688091 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:33 crc kubenswrapper[4754]: E1011 03:06:33.688397 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:33 crc kubenswrapper[4754]: E1011 03:06:33.688525 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:35.688492977 +0000 UTC m=+43.247437802 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.701527 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.723829 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740125 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740148 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.740195 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.757889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.778723 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.843034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.843086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.843098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.843116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.843128 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.946395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.946441 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.946449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.946464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:33 crc kubenswrapper[4754]: I1011 03:06:33.946476 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:33Z","lastTransitionTime":"2025-10-11T03:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.049904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.050006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.050019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.050035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.050045 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.083499 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.083601 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:34 crc kubenswrapper[4754]: E1011 03:06:34.083821 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:34 crc kubenswrapper[4754]: E1011 03:06:34.083950 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.083626 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:34 crc kubenswrapper[4754]: E1011 03:06:34.084454 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.153669 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.153714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.153726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.153744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.153756 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.257558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.257618 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.257632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.257652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.257666 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.360748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.360797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.360809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.360830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.360845 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.464234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.464336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.464355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.464377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.464681 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.567628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.567697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.567720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.567752 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.567776 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.670927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.671070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.671096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.671133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.671160 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.774211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.774289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.774314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.774343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.774363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.877348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.877423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.877445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.877472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.877492 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.981223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.981283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.981301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.981327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:34 crc kubenswrapper[4754]: I1011 03:06:34.981346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:34Z","lastTransitionTime":"2025-10-11T03:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.083429 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:35 crc kubenswrapper[4754]: E1011 03:06:35.083712 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.084552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.084593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.084621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.084648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.084670 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.188336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.188414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.188437 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.188466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.188487 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.291839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.291930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.291959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.292037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.292062 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.394682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.394727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.394735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.394750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.394760 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.499060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.499132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.499155 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.499185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.499208 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.604491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.604575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.604592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.604621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.604642 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.714024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.714186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.714328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.714410 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.714436 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.715350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:35 crc kubenswrapper[4754]: E1011 03:06:35.716488 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:35 crc kubenswrapper[4754]: E1011 03:06:35.716707 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:39.716666945 +0000 UTC m=+47.275611890 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.820804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.821129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.821155 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.821188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.821206 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.924159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.924221 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.924244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.924275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:35 crc kubenswrapper[4754]: I1011 03:06:35.924295 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:35Z","lastTransitionTime":"2025-10-11T03:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.027833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.027898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.027915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.027942 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.027988 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.083226 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.083272 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.083355 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:36 crc kubenswrapper[4754]: E1011 03:06:36.083459 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:36 crc kubenswrapper[4754]: E1011 03:06:36.083587 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:36 crc kubenswrapper[4754]: E1011 03:06:36.083742 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.131620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.131694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.131712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.131742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.131763 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.235034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.235119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.235141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.235172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.235194 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.338324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.338380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.338401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.338427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.338444 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.441988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.442062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.442081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.442113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.442136 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.546799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.546873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.546894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.546924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.546947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.651157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.651212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.651229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.651257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.651276 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.754434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.754492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.754507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.754551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.754566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.857639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.857696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.857716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.857740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.857759 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.961151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.961202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.961216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.961237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:36 crc kubenswrapper[4754]: I1011 03:06:36.961253 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:36Z","lastTransitionTime":"2025-10-11T03:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.065066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.065129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.065144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.065170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.065191 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.083436 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:37 crc kubenswrapper[4754]: E1011 03:06:37.083660 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.168107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.168200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.168225 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.168264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.168288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.272563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.272632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.272654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.272685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.272710 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.376571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.376629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.376648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.376673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.376691 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.480718 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.480803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.480825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.480854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.480877 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.585038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.585130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.585157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.585196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.585225 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.688272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.688320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.688330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.688345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.688354 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.791115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.791191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.791212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.791246 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.791266 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.895110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.895185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.895203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.895232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:37 crc kubenswrapper[4754]: I1011 03:06:37.895255 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:37Z","lastTransitionTime":"2025-10-11T03:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.000146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.000214 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.000232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.000259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.000278 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.083609 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.083651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.083714 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:38 crc kubenswrapper[4754]: E1011 03:06:38.083810 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:38 crc kubenswrapper[4754]: E1011 03:06:38.083988 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:38 crc kubenswrapper[4754]: E1011 03:06:38.084263 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.103159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.103202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.103220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.103244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.103262 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.206149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.206215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.206238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.206265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.206290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.310818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.310875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.310893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.310921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.310940 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.413296 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.413364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.413384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.413418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.413439 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.516381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.516449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.516470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.516496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.516518 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.620644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.620731 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.620756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.620788 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.620813 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.724956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.725099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.725144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.725186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.725215 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.828313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.828402 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.828430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.828470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.828496 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.931915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.931986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.932002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.932024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.932036 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.997188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.997263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.997282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.997349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:38 crc kubenswrapper[4754]: I1011 03:06:38.997378 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:38Z","lastTransitionTime":"2025-10-11T03:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.018335 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.024938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.025089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.025151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.025180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.025200 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.045202 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.050584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.050655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.050679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.050712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.050734 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.072549 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.077540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.077620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.077643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.077668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.077727 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.082888 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.083120 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.117450 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.128973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.129015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.129046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.129062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.129072 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.177921 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.178069 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.180070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.180130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.180141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.180159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.180170 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.282902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.282994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.283014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.283061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.283082 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.386404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.386444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.386460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.386481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.386496 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.489489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.489581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.489602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.489633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.489656 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.593401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.593480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.593496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.593520 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.593535 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.696649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.696685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.696695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.696711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.696720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.770585 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.770800 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:39 crc kubenswrapper[4754]: E1011 03:06:39.770919 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:06:47.770892707 +0000 UTC m=+55.329837482 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.799164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.799247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.799268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.799301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.799321 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.902279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.902338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.902358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.902386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:39 crc kubenswrapper[4754]: I1011 03:06:39.902406 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:39Z","lastTransitionTime":"2025-10-11T03:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.006795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.006856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.006874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.006899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.006917 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.083282 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.083350 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.083330 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:40 crc kubenswrapper[4754]: E1011 03:06:40.083521 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:40 crc kubenswrapper[4754]: E1011 03:06:40.083720 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:40 crc kubenswrapper[4754]: E1011 03:06:40.083902 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.110486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.110551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.110571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.110599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.110640 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.213566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.213608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.213620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.213638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.213650 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.316485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.316542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.316555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.316577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.316592 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.419325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.419403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.419422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.419450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.419466 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.522463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.522504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.522513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.522530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.522541 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.625131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.625187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.625200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.625222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.625239 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.728635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.728682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.728691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.728715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.728728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.832452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.832507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.832517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.832534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.832544 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.935639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.935711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.935730 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.935762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:40 crc kubenswrapper[4754]: I1011 03:06:40.935781 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:40Z","lastTransitionTime":"2025-10-11T03:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.038886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.038935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.038956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.039012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.039030 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.082717 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:41 crc kubenswrapper[4754]: E1011 03:06:41.082880 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.084004 4754 scope.go:117] "RemoveContainer" containerID="ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.142398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.142728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.142747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.142774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.142795 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.246450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.246507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.246524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.246548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.246566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.350843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.350902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.350921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.350946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.351000 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.417479 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/1.log" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.421872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.422633 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.449277 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.458279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.458353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.458377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.458409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.458430 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.477430 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.503563 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.526872 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.553438 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.561898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.561954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.562004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.562034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.562054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.613263 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.635780 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.663717 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.664288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.664332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.664345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.664364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.664378 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.678024 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.690915 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.704093 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.715814 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.726226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.742450 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.757842 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.767270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.767338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.767349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.767370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.767382 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.773653 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.789349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.870880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.870938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.870954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.870998 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.871016 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.974232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.974306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.974322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.974348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:41 crc kubenswrapper[4754]: I1011 03:06:41.974365 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:41Z","lastTransitionTime":"2025-10-11T03:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.078370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.078421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.078434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.078453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.078464 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.082580 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.082635 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.082679 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:42 crc kubenswrapper[4754]: E1011 03:06:42.082806 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:42 crc kubenswrapper[4754]: E1011 03:06:42.082904 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:42 crc kubenswrapper[4754]: E1011 03:06:42.083195 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.181826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.181880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.181895 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.181916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.181930 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.285267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.285322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.285339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.285365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.285381 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.388849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.388919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.388937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.388995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.389025 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.428487 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/2.log" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.429638 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/1.log" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.433107 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" exitCode=1 Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.433168 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.433225 4754 scope.go:117] "RemoveContainer" containerID="ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.433814 4754 scope.go:117] "RemoveContainer" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" Oct 11 03:06:42 crc kubenswrapper[4754]: E1011 03:06:42.433994 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.449634 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.476140 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.493607 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.493683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.493703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.493735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.493757 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.496722 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.518920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.536437 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.572735 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.596442 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.597728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.597810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.597830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.597862 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.597951 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.618830 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.641436 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.659506 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.685513 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.701587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.701632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.701641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.701658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.701669 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.702863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.718387 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.738603 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.756727 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.776144 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.794266 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.804611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.804700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.804727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.804765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.804794 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.908216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.908300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.908310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.908345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:42 crc kubenswrapper[4754]: I1011 03:06:42.908358 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:42Z","lastTransitionTime":"2025-10-11T03:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.012152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.012399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.012429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.012468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.012498 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.083780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.084091 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.107487 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.116400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.116489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.116507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.116538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.116561 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.133736 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.163050 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.184310 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.202006 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.219616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.219679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.219703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.219740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.219766 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.223905 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.241004 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.266718 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.289883 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.315810 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.323612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.323659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.323677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.323700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.323720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.339893 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.379850 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.400468 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.418616 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.426495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.426563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.426588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.426624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.426648 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.439335 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/2.log" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.444205 4754 scope.go:117] "RemoveContainer" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.444373 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.448298 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.483932 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba397182e5520f10fa6141e036e761debafe61f4ae4292c19ed8da5479130cad\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:28Z\\\",\\\"message\\\":\\\"1011 03:06:28.171646 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-machine-config-operator/machine-config-controller for endpointslice openshift-machine-config-operator/machine-config-controller-5lh94 as it is not a known egress service\\\\nI1011 03:06:28.171669 6196 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/certified-operators for endpointslice openshift-marketplace/certified-operators-7qh9v as it is not a known egress service\\\\nI1011 03:06:28.171752 6196 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.171792 6196 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.171852 6196 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:28.172029 6196 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:28.172334 6196 ovnkube.go:599] Stopped ovnkube\\\\nI1011 03:06:28.172384 6196 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1011 03:06:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.503245 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.520402 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.530587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.530649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.530661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.530685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.530703 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.535131 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.553269 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.574279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.590374 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.604567 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.628719 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.634317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.634412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.634433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.634502 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.634523 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.650822 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.672359 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.698851 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.721075 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.726182 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.726395 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:07:15.726361154 +0000 UTC m=+83.285305949 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.726512 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.726572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.726678 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.726721 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:15.726711583 +0000 UTC m=+83.285656388 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.727106 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.727215 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:15.727187916 +0000 UTC m=+83.286132751 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.737547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.737613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.737630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.737653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.737670 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.739819 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.756610 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.771631 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.787926 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.810104 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.828281 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.828355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828514 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828603 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828621 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828672 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828736 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828759 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828691 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:15.828671659 +0000 UTC m=+83.387616454 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:43 crc kubenswrapper[4754]: E1011 03:06:43.828880 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:15.828847834 +0000 UTC m=+83.387792799 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.829399 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.841066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.841272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.841421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.841558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.841692 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.944856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.944940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.945029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.945063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:43 crc kubenswrapper[4754]: I1011 03:06:43.945088 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:43Z","lastTransitionTime":"2025-10-11T03:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.048497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.048900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.049072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.049211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.049382 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.083058 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.083187 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.083445 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:44 crc kubenswrapper[4754]: E1011 03:06:44.083666 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:44 crc kubenswrapper[4754]: E1011 03:06:44.083931 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:44 crc kubenswrapper[4754]: E1011 03:06:44.084092 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.153338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.153811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.153999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.154164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.154299 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.258439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.259072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.259415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.259737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.260083 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.363799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.364204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.364408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.364605 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.364789 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.468003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.468268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.468328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.468367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.468390 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.571459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.571562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.571589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.571628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.571657 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.675772 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.675817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.675829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.675851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.675868 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.779413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.779462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.779478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.779496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.779514 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.882739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.882795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.882824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.882859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.882876 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.986754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.986810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.986824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.986843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:44 crc kubenswrapper[4754]: I1011 03:06:44.986857 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:44Z","lastTransitionTime":"2025-10-11T03:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.083194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:45 crc kubenswrapper[4754]: E1011 03:06:45.083427 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.093835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.093900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.093919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.093947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.094000 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.197223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.197286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.197303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.197328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.197345 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.300407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.300794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.300928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.301105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.301268 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.404254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.404320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.404334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.404352 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.404363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.509955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.510047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.510093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.510137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.510213 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.613430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.613471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.613482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.613497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.613509 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.717245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.717294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.717306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.717324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.717336 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.820612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.820680 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.820696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.820719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.820738 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.923632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.923695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.923773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.923810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:45 crc kubenswrapper[4754]: I1011 03:06:45.923831 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:45Z","lastTransitionTime":"2025-10-11T03:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.027229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.027326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.027352 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.027384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.027408 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.083424 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.083546 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.083620 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:46 crc kubenswrapper[4754]: E1011 03:06:46.083801 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:46 crc kubenswrapper[4754]: E1011 03:06:46.083941 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:46 crc kubenswrapper[4754]: E1011 03:06:46.084104 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.131099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.131190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.131209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.131233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.131282 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.235042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.235111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.235128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.235153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.235175 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.338095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.338162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.338178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.338210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.338252 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.441945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.442035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.442055 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.442082 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.442104 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.545573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.545680 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.545698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.545729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.545747 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.649144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.649229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.649444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.649471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.649490 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.753105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.753181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.753199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.753405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.753424 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.856451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.856538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.856556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.856610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.856626 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.959535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.959577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.959589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.959606 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:46 crc kubenswrapper[4754]: I1011 03:06:46.959619 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:46Z","lastTransitionTime":"2025-10-11T03:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.063869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.063929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.063944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.064012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.064027 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.083405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:47 crc kubenswrapper[4754]: E1011 03:06:47.083600 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.170034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.170134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.170161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.170204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.170241 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.274401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.274465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.274479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.274500 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.274524 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.379208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.379277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.379291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.379316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.379328 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.482258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.482324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.482341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.482367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.482387 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.585412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.585501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.585528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.585560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.585578 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.687870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.687911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.687923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.687945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.687983 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.771801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:47 crc kubenswrapper[4754]: E1011 03:06:47.771933 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:47 crc kubenswrapper[4754]: E1011 03:06:47.772022 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:03.772005271 +0000 UTC m=+71.330950056 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.790272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.790304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.790312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.790326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.790337 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.892854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.892908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.892921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.892937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.892948 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.996174 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.996251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.996270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.996302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:47 crc kubenswrapper[4754]: I1011 03:06:47.996325 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:47Z","lastTransitionTime":"2025-10-11T03:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.083413 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.083518 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.083426 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:48 crc kubenswrapper[4754]: E1011 03:06:48.083657 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:48 crc kubenswrapper[4754]: E1011 03:06:48.083790 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:48 crc kubenswrapper[4754]: E1011 03:06:48.083917 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.099694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.099775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.099792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.099823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.099844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.203280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.203355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.203379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.203404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.203421 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.306280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.306365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.306386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.306424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.306447 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.410016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.410115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.410141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.410180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.410205 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.517382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.517447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.517470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.517499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.517523 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.620805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.620872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.620895 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.620923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.620940 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.723441 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.723502 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.723522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.723549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.723566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.826999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.827100 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.827119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.827153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.827175 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.930358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.930435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.930465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.930493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:48 crc kubenswrapper[4754]: I1011 03:06:48.930513 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:48Z","lastTransitionTime":"2025-10-11T03:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.033622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.033682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.033698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.033726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.033746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.051651 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.061605 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.069714 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.083109 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.083231 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.088387 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.114084 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.134886 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.137340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.137411 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.137442 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.137475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.137499 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.146896 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.157263 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.182316 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.196458 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.207693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.227930 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.236709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.236768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.236778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.236796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.236809 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.241549 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.247266 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.250811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.250835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.250845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.250860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.250869 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.258376 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.261438 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.265067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.265126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.265143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.265170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.265191 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.273035 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.279371 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.283093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.283174 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.283189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.283213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.283228 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.290384 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.298989 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.301568 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.303580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.303626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.303642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.303663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.303677 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.314455 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.315636 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: E1011 03:06:49.315761 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.317325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.317365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.317378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.317396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.317411 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.327161 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:49Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.420541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.420585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.420594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.420610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.420621 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.523799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.523867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.523894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.523925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.524031 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.626648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.626697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.626715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.626736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.626753 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.729816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.729896 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.729915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.729945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.730101 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.833181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.833254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.833272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.833297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.833320 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.937856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.937936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.937956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.938026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:49 crc kubenswrapper[4754]: I1011 03:06:49.938052 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:49Z","lastTransitionTime":"2025-10-11T03:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.041882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.041949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.042001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.042027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.042051 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.083633 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:50 crc kubenswrapper[4754]: E1011 03:06:50.083808 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.083900 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.084008 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:50 crc kubenswrapper[4754]: E1011 03:06:50.084251 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:50 crc kubenswrapper[4754]: E1011 03:06:50.084510 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.144594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.144709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.144728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.144757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.144778 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.247576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.247625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.247641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.247668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.247687 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.352252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.352323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.352344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.352373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.352401 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.455266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.455490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.455511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.455530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.455544 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.558832 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.558871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.558881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.558898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.558908 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.664626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.664692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.664711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.664740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.664761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.768507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.768578 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.768596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.768630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.768660 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.872152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.872234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.872263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.872301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.872328 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.976423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.976548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.976574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.976604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:50 crc kubenswrapper[4754]: I1011 03:06:50.976633 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:50Z","lastTransitionTime":"2025-10-11T03:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.080851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.080917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.080941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.081003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.081031 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.083543 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:51 crc kubenswrapper[4754]: E1011 03:06:51.083727 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.184829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.184905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.184931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.184993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.185033 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.288484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.288571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.288594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.288628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.288652 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.392318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.392371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.392408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.392433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.392449 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.495676 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.495756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.495774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.495800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.495819 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.599562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.599621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.599639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.599671 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.599689 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.703147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.703307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.703345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.703375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.703395 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.807902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.808016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.808039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.808070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.808089 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.911258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.911339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.911361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.911395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:51 crc kubenswrapper[4754]: I1011 03:06:51.911415 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:51Z","lastTransitionTime":"2025-10-11T03:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.014690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.014745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.014757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.014778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.014792 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.082846 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:52 crc kubenswrapper[4754]: E1011 03:06:52.083027 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.082871 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:52 crc kubenswrapper[4754]: E1011 03:06:52.083141 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.082846 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:52 crc kubenswrapper[4754]: E1011 03:06:52.083249 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.117426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.117463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.117477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.117494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.117510 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.220535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.220617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.220662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.220695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.220722 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.323740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.323770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.323778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.323794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.323806 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.426440 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.426481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.426493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.426507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.426516 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.529106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.529161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.529180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.529204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.529226 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.634456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.634550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.634577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.634613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.634648 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.738802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.738887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.738908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.738938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.738995 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.842036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.842103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.842127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.842158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.842182 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.944766 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.944856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.944874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.945399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:52 crc kubenswrapper[4754]: I1011 03:06:52.945458 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:52Z","lastTransitionTime":"2025-10-11T03:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.048068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.048142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.048163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.048188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.048208 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.083343 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:53 crc kubenswrapper[4754]: E1011 03:06:53.083572 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.106080 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.126118 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.142891 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.151187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.151209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.151217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.151232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.151241 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.177413 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.198053 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.214516 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.233943 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.249546 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.257310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.257384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.257405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.257435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.257451 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.268766 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.286942 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.304877 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.321336 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.341306 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.355863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.360123 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.360208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.360247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.360268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.360280 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.372541 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.390434 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.404213 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.419792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.462930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.463046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.463058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.463074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.463084 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.565543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.565592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.565605 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.565622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.565633 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.668090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.668169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.668181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.668200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.668212 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.771297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.771356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.771391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.771429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.771454 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.874609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.874668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.874679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.874697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.874710 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.977292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.977367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.977390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.977425 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:53 crc kubenswrapper[4754]: I1011 03:06:53.977445 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:53Z","lastTransitionTime":"2025-10-11T03:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.080644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.080717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.080742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.080782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.080806 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.083196 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.083278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:54 crc kubenswrapper[4754]: E1011 03:06:54.083405 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.083438 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:54 crc kubenswrapper[4754]: E1011 03:06:54.083618 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:54 crc kubenswrapper[4754]: E1011 03:06:54.083743 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.184025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.184063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.184075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.184091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.184107 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.287647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.287709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.287728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.287751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.287767 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.390919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.390991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.391008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.391027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.391041 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.494193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.494291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.494314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.494335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.494381 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.597638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.597699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.597717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.597744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.597766 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.700655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.700708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.700721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.700740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.700754 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.804019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.804098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.804117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.804573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.804640 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.908395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.908496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.908516 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.908543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:54 crc kubenswrapper[4754]: I1011 03:06:54.908562 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:54Z","lastTransitionTime":"2025-10-11T03:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.011799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.011855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.011873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.011897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.011914 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.083376 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:55 crc kubenswrapper[4754]: E1011 03:06:55.083592 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.114630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.114720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.114745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.114774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.114797 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.217706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.217762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.217775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.217798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.217814 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.321498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.321562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.321573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.321596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.321608 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.423675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.423750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.423776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.423807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.423830 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.526188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.526304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.526315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.526335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.526347 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.629845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.629918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.629930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.629994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.630011 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.732586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.732631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.732639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.732654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.732664 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.836229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.836297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.836315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.836347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.836366 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.938613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.938657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.938668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.938683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:55 crc kubenswrapper[4754]: I1011 03:06:55.938720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:55Z","lastTransitionTime":"2025-10-11T03:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.041901 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.041977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.041987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.042004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.042014 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.082756 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.082815 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.082893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:56 crc kubenswrapper[4754]: E1011 03:06:56.083071 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:56 crc kubenswrapper[4754]: E1011 03:06:56.083221 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:56 crc kubenswrapper[4754]: E1011 03:06:56.083314 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.084153 4754 scope.go:117] "RemoveContainer" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" Oct 11 03:06:56 crc kubenswrapper[4754]: E1011 03:06:56.084383 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.144726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.144790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.144803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.144823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.144841 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.248374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.248423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.248447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.248472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.248510 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.351483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.351536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.351547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.351566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.351582 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.454853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.454937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.454957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.455017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.455037 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.558798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.558849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.558860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.558878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.558892 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.662039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.662105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.662117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.662138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.662149 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.765337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.765382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.765392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.765409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.765422 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.869932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.870014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.870029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.870050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.870070 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.972880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.972959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.973016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.973045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:56 crc kubenswrapper[4754]: I1011 03:06:56.973064 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:56Z","lastTransitionTime":"2025-10-11T03:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.076098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.076155 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.076167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.076189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.076205 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.083660 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:57 crc kubenswrapper[4754]: E1011 03:06:57.083840 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.178431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.178477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.178493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.178513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.178529 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.281635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.281685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.281706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.281727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.281739 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.383835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.383890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.383902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.383920 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.383933 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.486781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.486841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.486855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.486903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.486917 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.589306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.589366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.589384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.589409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.589427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.691928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.692026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.692040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.692068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.692082 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.794626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.794688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.794705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.794730 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.794747 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.898427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.898564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.898586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.898649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:57 crc kubenswrapper[4754]: I1011 03:06:57.898672 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:57Z","lastTransitionTime":"2025-10-11T03:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.001060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.001108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.001119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.001139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.001159 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.083029 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.083108 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:06:58 crc kubenswrapper[4754]: E1011 03:06:58.083193 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.083295 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:06:58 crc kubenswrapper[4754]: E1011 03:06:58.083483 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:06:58 crc kubenswrapper[4754]: E1011 03:06:58.083757 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.102811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.102844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.102853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.102869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.102881 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.205841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.205921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.205940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.206005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.206029 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.308890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.308932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.308941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.308957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.308980 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.412166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.412243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.412267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.412296 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.412316 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.519716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.519794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.519820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.519853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.519876 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.622451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.622489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.622500 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.622517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.622528 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.724955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.725020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.725036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.725078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.725092 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.827982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.828016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.828024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.828038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.828047 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.930886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.930935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.931088 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.931106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:58 crc kubenswrapper[4754]: I1011 03:06:58.931118 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:58Z","lastTransitionTime":"2025-10-11T03:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.034020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.034069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.034096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.034126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.034138 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.084209 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.084328 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.136950 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.137022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.137033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.137055 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.137069 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.239820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.239863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.239874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.239891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.239902 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.343256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.343317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.343327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.343351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.343363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.446657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.446732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.446753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.446793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.446820 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.518404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.518465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.518486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.518512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.518532 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.539125 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:59Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.545549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.545616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.545647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.545681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.545709 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.560215 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:59Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.564916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.565020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.565050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.565081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.565107 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.582383 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:59Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.588094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.588137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.588158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.588183 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.588202 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.608568 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:59Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.613000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.613037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.613049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.613069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.613087 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.633908 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:06:59Z is after 2025-08-24T17:21:41Z" Oct 11 03:06:59 crc kubenswrapper[4754]: E1011 03:06:59.634390 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.636477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.636524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.636535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.636555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.636566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.742237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.742340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.742426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.742468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.742494 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.846385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.846439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.846477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.846499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.846512 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.951520 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.952834 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.953255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.953324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:06:59 crc kubenswrapper[4754]: I1011 03:06:59.953353 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:06:59Z","lastTransitionTime":"2025-10-11T03:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.056676 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.056725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.056740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.056760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.056775 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.083468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.083525 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.083527 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:00 crc kubenswrapper[4754]: E1011 03:07:00.083621 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:00 crc kubenswrapper[4754]: E1011 03:07:00.083714 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:00 crc kubenswrapper[4754]: E1011 03:07:00.083879 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.159820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.159884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.159900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.159921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.159936 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.263414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.263483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.263503 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.263535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.263555 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.366442 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.366496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.366509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.366532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.366545 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.469799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.469878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.469903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.469935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.469956 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.572485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.572539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.572559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.572583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.572599 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.675904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.675956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.675984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.676005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.676017 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.778852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.778899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.778909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.778926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.778939 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.882534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.882613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.882631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.882660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.882679 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.986626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.986683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.986700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.986726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:00 crc kubenswrapper[4754]: I1011 03:07:00.986744 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:00Z","lastTransitionTime":"2025-10-11T03:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.083509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:01 crc kubenswrapper[4754]: E1011 03:07:01.083686 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.090291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.090327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.090342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.090362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.090380 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.196342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.196400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.196412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.196432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.196446 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.299211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.299249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.299259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.299276 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.299288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.401919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.401981 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.401992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.402011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.402023 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.504230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.504259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.504267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.504280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.504290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.606850 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.606921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.606947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.607008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.607032 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.709602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.709649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.709665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.709687 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.709706 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.811779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.811854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.811874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.811906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.811937 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.915017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.915075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.915093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.915116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:01 crc kubenswrapper[4754]: I1011 03:07:01.915136 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:01Z","lastTransitionTime":"2025-10-11T03:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.018415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.018459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.018474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.018491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.018504 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.083491 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:02 crc kubenswrapper[4754]: E1011 03:07:02.083631 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.083511 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.083491 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:02 crc kubenswrapper[4754]: E1011 03:07:02.083699 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:02 crc kubenswrapper[4754]: E1011 03:07:02.084003 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.120789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.120861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.120879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.120909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.120930 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.226231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.226324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.226342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.226367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.226384 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.329844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.329894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.329905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.329925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.329939 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.432736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.432815 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.432836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.432868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.432888 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.536292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.536360 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.536380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.536412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.536436 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.639427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.639478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.639491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.639514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.639528 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.743274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.743336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.743350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.743367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.743380 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.846479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.846553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.846565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.846588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.846606 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.949163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.949230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.949294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.949325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:02 crc kubenswrapper[4754]: I1011 03:07:02.949346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:02Z","lastTransitionTime":"2025-10-11T03:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.053231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.053792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.053814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.053833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.053846 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.083441 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:03 crc kubenswrapper[4754]: E1011 03:07:03.083564 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.112445 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.130804 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.153555 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.156888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.156995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.157016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.157045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.157064 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.188906 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.212926 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.229802 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.254356 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.259546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.259622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.259641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.259669 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.259694 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.270737 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.292496 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.306916 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.322023 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.336536 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.356779 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.363086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.363164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.363194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.363233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.363264 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.377215 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.396238 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.414760 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.433725 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.455483 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.466239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.466282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.466293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.466312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.466323 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.569691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.569747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.569760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.569784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.569800 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.673166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.673342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.673372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.673407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.673431 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.776403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.776451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.776465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.776491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.776508 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.858956 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:03 crc kubenswrapper[4754]: E1011 03:07:03.859255 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:07:03 crc kubenswrapper[4754]: E1011 03:07:03.859372 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:07:35.859349114 +0000 UTC m=+103.418293909 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.882149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.882237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.882261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.882294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.882318 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.985399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.985446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.985459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.985478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:03 crc kubenswrapper[4754]: I1011 03:07:03.985496 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:03Z","lastTransitionTime":"2025-10-11T03:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.082644 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.082737 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:04 crc kubenswrapper[4754]: E1011 03:07:04.082797 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.082666 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:04 crc kubenswrapper[4754]: E1011 03:07:04.082863 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:04 crc kubenswrapper[4754]: E1011 03:07:04.082997 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.088816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.088876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.088888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.088909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.088922 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.191752 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.191814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.191871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.191901 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.191929 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.295326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.295371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.295384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.295401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.295413 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.399213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.399285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.399304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.399337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.399356 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.502552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.502593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.502602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.502617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.502628 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.605735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.605768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.605779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.605795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.605806 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.709293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.709380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.709406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.709448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.709475 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.812955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.812999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.813008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.813023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.813033 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.916057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.916112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.916125 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.916149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:04 crc kubenswrapper[4754]: I1011 03:07:04.916166 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:04Z","lastTransitionTime":"2025-10-11T03:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.019918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.019982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.019994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.020013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.020028 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.083357 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:05 crc kubenswrapper[4754]: E1011 03:07:05.083497 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.123193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.123251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.123268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.123296 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.123317 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.233180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.233284 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.233315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.233354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.233380 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.336989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.337042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.337056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.337079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.337092 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.441268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.441330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.441346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.441370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.441386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.543382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.543439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.543457 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.543485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.543505 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.646857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.646897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.646910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.646930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.646943 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.749012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.749044 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.749052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.749066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.749077 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.851371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.851428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.851445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.851471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.851493 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.955121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.955192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.955210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.955234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:05 crc kubenswrapper[4754]: I1011 03:07:05.955254 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:05Z","lastTransitionTime":"2025-10-11T03:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.058021 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.058090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.058107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.058133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.058150 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.083286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.083340 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.083287 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:06 crc kubenswrapper[4754]: E1011 03:07:06.083421 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:06 crc kubenswrapper[4754]: E1011 03:07:06.083519 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:06 crc kubenswrapper[4754]: E1011 03:07:06.083732 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.161247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.161315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.161332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.161359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.161379 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.265774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.265853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.265878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.265911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.265934 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.369351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.369400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.369416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.369455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.369471 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.472167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.472220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.472232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.472251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.472262 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.519141 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/0.log" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.519213 4754 generic.go:334] "Generic (PLEG): container finished" podID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" containerID="acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b" exitCode=1 Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.519253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerDied","Data":"acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.519838 4754 scope.go:117] "RemoveContainer" containerID="acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.541580 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.562685 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.575593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.575638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.575653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.575674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.575687 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.580484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.598841 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.615195 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.631679 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.661124 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.674421 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.680094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.680129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.680145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.680165 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.680179 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.693427 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.709731 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.730082 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.751955 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.782655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.782723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.782736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.782755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.782769 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.789996 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.814505 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.839227 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.866311 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.884414 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.886444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.886492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.886505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.886537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.886549 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.901509 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:06Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.990283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.990320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.990330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.990350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:06 crc kubenswrapper[4754]: I1011 03:07:06.990362 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:06Z","lastTransitionTime":"2025-10-11T03:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.083770 4754 scope.go:117] "RemoveContainer" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.084298 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:07 crc kubenswrapper[4754]: E1011 03:07:07.084421 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.093780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.093829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.093854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.093886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.093911 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.111069 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.197022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.197345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.197510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.198166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.198346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.302162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.302222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.302240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.302267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.302284 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.435598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.435654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.435668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.435690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.435705 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.529572 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/2.log" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.534890 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.535750 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.540870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.540911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.540929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.540955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.541001 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.542656 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/0.log" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.543686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerStarted","Data":"5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.551659 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.576422 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.600859 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.617723 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.640732 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.643864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.643926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.643946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.644022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.644044 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.655282 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.668256 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.687242 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.702655 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.725346 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.742560 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.746589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.746727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.746819 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.746892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.746983 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.762842 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.777768 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.794298 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.811324 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.824168 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.834641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.846543 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.849549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.849656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.849724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.849807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.849883 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.856688 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.868030 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.881330 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.890918 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.904565 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.916362 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.933786 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.952099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.952131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.952142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.952158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.952168 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:07Z","lastTransitionTime":"2025-10-11T03:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.954837 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.971226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:07 crc kubenswrapper[4754]: I1011 03:07:07.983856 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.001476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:07Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.013735 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.029337 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.042259 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.052927 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.054771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.054802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.054812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.054827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.054837 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.071628 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.083347 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.083461 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:08 crc kubenswrapper[4754]: E1011 03:07:08.083587 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.083644 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:08 crc kubenswrapper[4754]: E1011 03:07:08.083741 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:08 crc kubenswrapper[4754]: E1011 03:07:08.083776 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.085531 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.101863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.113750 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.125589 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.157647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.157716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.157736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.157765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.157785 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.261665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.261727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.261739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.261758 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.261774 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.365777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.365843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.365861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.365888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.365910 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.470057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.470109 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.470126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.470151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.470169 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.548731 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/3.log" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.549704 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/2.log" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.552377 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" exitCode=1 Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.552420 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.552463 4754 scope.go:117] "RemoveContainer" containerID="ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.553472 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:08 crc kubenswrapper[4754]: E1011 03:07:08.553730 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.564444 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.572885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.572927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.572937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.572954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.572979 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.580110 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.592312 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.608844 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.625695 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.642053 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.658326 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.673166 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.675881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.675924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.675936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.675954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.675992 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.690473 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.702898 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.717748 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.730997 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.742399 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.759102 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea13bea42cf3facf7320ed2be11cc1716a24011ed67e644280280dd62439c22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:06:42Z\\\",\\\"message\\\":\\\"ode (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:06:42.058646 6410 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1011 03:06:42.058672 6410 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1011 03:06:42.058696 6410 factory.go:656] Stopping watch factory\\\\nI1011 03:06:42.058700 6410 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058723 6410 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1011 03:06:42.058790 6410 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.058708 6410 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1011 03:06:42.058647 6410 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:06:42.059068 6410 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:08Z\\\",\\\"message\\\":\\\"x2sf\\\\nI1011 03:07:08.034497 6762 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1011 03:07:08.034213 6762 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1011 03:07:08.034576 6762 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034723 6762 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-2nlvv in node crc\\\\nI1011 03:07:08.034729 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-2nlvv after 0 failed attempt(s)\\\\nI1011 03:07:08.034734 6762 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034484 6762 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1011 03:07:08.034749 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nF1011 03:07:08.034752 6762 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fai\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.770849 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.778175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.778220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.778237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.778259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.778275 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.786782 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.807535 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.821401 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.834619 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:08Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.880598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.880644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.880655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.880672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.880685 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.983432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.983481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.983491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.983508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:08 crc kubenswrapper[4754]: I1011 03:07:08.983518 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:08Z","lastTransitionTime":"2025-10-11T03:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.083135 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.083317 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.085795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.085839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.085849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.085864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.085874 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.188739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.188812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.188829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.188855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.188876 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.292737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.292801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.292823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.292851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.292871 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.395601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.395669 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.395692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.395720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.395745 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.499564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.499618 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.499642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.499664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.499678 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.559392 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/3.log" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.564661 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.565035 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.586636 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.603430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.603493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.603518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.603544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.603566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.606794 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.624491 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.647429 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.665209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.665277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.665302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.665333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.665357 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.669331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.690066 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.697468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.697563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.697583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.697616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.697635 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.709684 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.718693 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.726112 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.726907 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.727049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.727071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.727098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.727118 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.744959 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.750904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.751022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.751107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.751286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.751330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.753228 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.766469 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.772330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.772364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.772378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.772397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.772411 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.773095 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.794044 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.797944 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: E1011 03:07:09.798275 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.800902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.800941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.800954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.801009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.801025 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.809633 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.833169 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:08Z\\\",\\\"message\\\":\\\"x2sf\\\\nI1011 03:07:08.034497 6762 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1011 03:07:08.034213 6762 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1011 03:07:08.034576 6762 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034723 6762 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-2nlvv in node crc\\\\nI1011 03:07:08.034729 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-2nlvv after 0 failed attempt(s)\\\\nI1011 03:07:08.034734 6762 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034484 6762 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1011 03:07:08.034749 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nF1011 03:07:08.034752 6762 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fai\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.850296 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.870046 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.887596 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.905114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.905191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.905207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.905263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.905281 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:09Z","lastTransitionTime":"2025-10-11T03:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.909278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.927819 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.944830 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:09 crc kubenswrapper[4754]: I1011 03:07:09.964553 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.008008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.008084 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.008095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.008112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.008121 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.083651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:10 crc kubenswrapper[4754]: E1011 03:07:10.083773 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.083669 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:10 crc kubenswrapper[4754]: E1011 03:07:10.083845 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.083667 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:10 crc kubenswrapper[4754]: E1011 03:07:10.083932 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.110786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.110829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.110842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.110860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.110874 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.214308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.214357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.214370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.214387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.214397 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.318381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.318471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.318489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.318536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.318557 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.421288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.421376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.421398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.421429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.421450 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.524457 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.524550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.524574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.524607 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.524632 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.627818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.627873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.627884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.627902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.628005 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.731128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.731184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.731196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.731218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.731232 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.834181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.834218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.834226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.834243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.834254 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.937125 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.937198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.937221 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.937253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:10 crc kubenswrapper[4754]: I1011 03:07:10.937272 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:10Z","lastTransitionTime":"2025-10-11T03:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.040058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.040094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.040105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.040120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.040136 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.083558 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:11 crc kubenswrapper[4754]: E1011 03:07:11.083712 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.143141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.143178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.143187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.143201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.143211 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.245941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.246074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.246095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.246130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.246151 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.349362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.349403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.349413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.349428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.349438 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.452121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.452171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.452184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.452203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.452215 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.555191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.555239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.555251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.555272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.555284 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.658482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.658549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.658569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.658595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.658616 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.762146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.762224 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.762243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.762270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.762289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.866427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.866492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.866510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.866535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.866555 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.969868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.969925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.969943 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.969992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:11 crc kubenswrapper[4754]: I1011 03:07:11.970011 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:11Z","lastTransitionTime":"2025-10-11T03:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.074030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.074072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.074082 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.074097 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.074108 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.082861 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.083009 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:12 crc kubenswrapper[4754]: E1011 03:07:12.083087 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.083207 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:12 crc kubenswrapper[4754]: E1011 03:07:12.083408 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:12 crc kubenswrapper[4754]: E1011 03:07:12.083785 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.177408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.177473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.177492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.177518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.177542 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.282803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.282869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.282887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.282914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.282932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.385779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.385839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.385855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.385882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.385899 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.493333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.493407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.493429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.493459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.493476 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.597244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.597308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.597330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.597358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.597381 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.700863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.700914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.700938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.700982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.700997 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.803236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.803295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.803322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.803349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.803367 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.906748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.906830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.906848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.906876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:12 crc kubenswrapper[4754]: I1011 03:07:12.906906 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:12Z","lastTransitionTime":"2025-10-11T03:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.010450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.011019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.011199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.011398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.011552 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.083594 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:13 crc kubenswrapper[4754]: E1011 03:07:13.083843 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.100939 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.116881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.116923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.116933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.116953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.116983 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.123621 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:08Z\\\",\\\"message\\\":\\\"x2sf\\\\nI1011 03:07:08.034497 6762 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1011 03:07:08.034213 6762 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1011 03:07:08.034576 6762 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034723 6762 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-2nlvv in node crc\\\\nI1011 03:07:08.034729 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-2nlvv after 0 failed attempt(s)\\\\nI1011 03:07:08.034734 6762 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034484 6762 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1011 03:07:08.034749 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nF1011 03:07:08.034752 6762 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fai\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.142606 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.155451 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.174787 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.192393 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.207816 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.220128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.220186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.220198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.220218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.220231 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.223711 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.237334 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.250009 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.268279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.288883 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.307317 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.322474 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.325518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.325547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.325556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.325574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.325586 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.338484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.360631 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.374160 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.393516 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.410617 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.428939 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.429028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.429048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.429078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.429096 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.531400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.532019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.532041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.532072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.532092 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.635707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.635781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.635799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.635829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.635848 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.739584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.739673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.739701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.739736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.739759 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.842300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.842336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.842346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.842364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.842375 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.945549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.945633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.945703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.945743 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:13 crc kubenswrapper[4754]: I1011 03:07:13.945773 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:13Z","lastTransitionTime":"2025-10-11T03:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.049953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.050011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.050023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.050040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.050054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.082782 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.082854 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:14 crc kubenswrapper[4754]: E1011 03:07:14.082936 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.083046 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:14 crc kubenswrapper[4754]: E1011 03:07:14.083065 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:14 crc kubenswrapper[4754]: E1011 03:07:14.083328 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.153366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.153471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.153490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.153518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.153538 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.257534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.257614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.257635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.257664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.257686 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.362514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.362603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.362640 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.362675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.362697 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.467514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.467596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.467620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.467656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.467698 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.571623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.571703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.571717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.571742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.571761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.675355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.675416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.675435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.675460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.675479 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.779226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.779273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.779286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.779303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.779315 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.883339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.883449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.883468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.883501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.883521 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.987581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.987666 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.987694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.987731 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:14 crc kubenswrapper[4754]: I1011 03:07:14.987755 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:14Z","lastTransitionTime":"2025-10-11T03:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.083604 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.083889 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.091301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.091375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.091396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.091423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.091441 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.195089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.195226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.195252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.195288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.195313 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.298661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.298707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.298719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.298739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.298749 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.403059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.403116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.403131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.403152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.403169 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.506586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.506635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.506644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.506660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.506674 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.609817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.609885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.609903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.609932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.609954 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.714734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.714806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.714828 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.714875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.715028 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.738503 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.738737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.738836 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.739122 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.739275 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.73924074 +0000 UTC m=+147.298185565 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.739369 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.739319712 +0000 UTC m=+147.298264557 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.739139 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.739516 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.739490437 +0000 UTC m=+147.298435512 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.817816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.817887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.817896 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.817913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.817926 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.841017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.841101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841175 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841207 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841218 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841283 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.841263073 +0000 UTC m=+147.400207858 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841383 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841427 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841455 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:07:15 crc kubenswrapper[4754]: E1011 03:07:15.841575 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.841533831 +0000 UTC m=+147.400478806 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.920563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.920699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.920724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.920754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:15 crc kubenswrapper[4754]: I1011 03:07:15.920777 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:15Z","lastTransitionTime":"2025-10-11T03:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.024560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.024635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.024660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.024692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.024715 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.082829 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.082857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.082857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:16 crc kubenswrapper[4754]: E1011 03:07:16.083115 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:16 crc kubenswrapper[4754]: E1011 03:07:16.083292 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:16 crc kubenswrapper[4754]: E1011 03:07:16.083413 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.128171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.128238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.128258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.128285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.128306 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.232443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.232496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.232512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.232537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.232558 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.336195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.336292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.336323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.336357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.336390 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.440186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.440279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.440302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.440335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.440360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.543778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.543886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.543916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.543953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.544025 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.648399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.648504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.648521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.648550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.648567 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.752042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.752093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.752111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.752136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.752154 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.856555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.856633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.856658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.856691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.856715 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.960255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.960320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.960340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.960367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:16 crc kubenswrapper[4754]: I1011 03:07:16.960389 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:16Z","lastTransitionTime":"2025-10-11T03:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.063473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.063565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.063586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.063620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.063643 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.082955 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:17 crc kubenswrapper[4754]: E1011 03:07:17.083233 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.167069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.167139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.167162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.167194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.167218 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.270733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.270814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.270824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.270843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.270854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.374106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.374160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.374172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.374191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.374207 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.479263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.479331 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.479347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.479370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.479385 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.583620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.583663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.583673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.583693 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.583703 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.686447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.686531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.686573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.686596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.686612 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.790436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.790522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.790535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.790555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.790574 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.893795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.893863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.893881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.893918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.893941 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.997780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.997846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.997865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.997894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:17 crc kubenswrapper[4754]: I1011 03:07:17.997914 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:17Z","lastTransitionTime":"2025-10-11T03:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.082982 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.083027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:18 crc kubenswrapper[4754]: E1011 03:07:18.083116 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.083251 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:18 crc kubenswrapper[4754]: E1011 03:07:18.083469 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:18 crc kubenswrapper[4754]: E1011 03:07:18.083655 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.101235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.101288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.101305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.101330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.101349 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.204586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.204665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.204688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.204717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.204738 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.308699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.308782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.308804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.308835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.308856 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.412064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.412128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.412145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.412176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.412196 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.516032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.516106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.516133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.516161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.516179 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.618818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.618870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.618891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.618921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.618947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.722242 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.722316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.722339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.722369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.722389 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.826362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.826431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.826449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.826485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.826508 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.933491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.933565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.933583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.933617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:18 crc kubenswrapper[4754]: I1011 03:07:18.933636 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:18Z","lastTransitionTime":"2025-10-11T03:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.038439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.038514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.038532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.038560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.038579 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.083990 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.084232 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.142764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.142835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.142852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.142877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.142897 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.246592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.246663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.246689 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.246725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.246749 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.357144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.357243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.357266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.357297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.357318 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.460233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.460294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.460311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.460335 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.460354 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.563313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.563390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.563415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.563451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.563476 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.666922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.667013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.667035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.667061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.667083 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.769996 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.770067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.770087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.770118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.770139 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.811089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.811147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.811164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.811194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.811212 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.831365 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.836131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.836200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.836224 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.836257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.836285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.859096 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.864160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.864207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.864217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.864234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.864247 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.885462 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.891141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.891168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.891179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.891194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.891205 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.913782 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.919481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.919524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.919537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.919560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.919578 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.935902 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f5c76452-3606-44cd-ab8e-b64535e75615\\\",\\\"systemUUID\\\":\\\"abd9917e-dee0-4bd6-812d-a0709f74ee02\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:19 crc kubenswrapper[4754]: E1011 03:07:19.936204 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.939645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.939747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.939768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.939805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:19 crc kubenswrapper[4754]: I1011 03:07:19.939836 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:19Z","lastTransitionTime":"2025-10-11T03:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.044701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.044751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.044769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.044796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.044807 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.082670 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.082670 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:20 crc kubenswrapper[4754]: E1011 03:07:20.082819 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:20 crc kubenswrapper[4754]: E1011 03:07:20.082917 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.082696 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:20 crc kubenswrapper[4754]: E1011 03:07:20.083003 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.148113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.148163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.148176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.148243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.148256 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.251280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.251337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.251349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.251372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.251386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.355166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.355247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.355266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.355295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.355313 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.458190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.458250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.458264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.458285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.458300 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.561391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.561436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.561449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.561663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.561679 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.664842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.664908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.664925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.664949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.664990 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.768231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.768288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.768306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.768330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.768348 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.870705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.870772 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.870804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.870836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.870857 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.974301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.974356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.974374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.974400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:20 crc kubenswrapper[4754]: I1011 03:07:20.974419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:20Z","lastTransitionTime":"2025-10-11T03:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.078727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.078800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.078827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.078864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.078885 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.085184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:21 crc kubenswrapper[4754]: E1011 03:07:21.085401 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.086563 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:21 crc kubenswrapper[4754]: E1011 03:07:21.086842 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.182376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.182472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.182497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.182535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.182560 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.286532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.286613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.286632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.286665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.286687 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.390588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.390666 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.390685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.390714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.390733 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.494655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.494728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.494753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.494789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.494818 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.598844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.598940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.599006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.599091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.599120 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.702357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.702439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.702460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.702488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.702507 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.806619 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.806696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.806719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.806751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.806773 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.917874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.918057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.918085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.918119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:21 crc kubenswrapper[4754]: I1011 03:07:21.918144 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:21Z","lastTransitionTime":"2025-10-11T03:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.021315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.021404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.021426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.021455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.021476 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.082654 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.082837 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.082910 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:22 crc kubenswrapper[4754]: E1011 03:07:22.083589 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:22 crc kubenswrapper[4754]: E1011 03:07:22.083704 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:22 crc kubenswrapper[4754]: E1011 03:07:22.084294 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.124845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.124901 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.124914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.124933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.124947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.228206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.228270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.228287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.228313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.228332 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.332002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.332079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.332102 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.332136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.332161 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.435206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.435281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.435301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.435333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.435354 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.539189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.539280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.539299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.539333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.539357 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.642641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.642692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.642707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.642730 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.642744 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.746095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.746195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.746219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.746253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.746272 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.850124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.850208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.850232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.850265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.850288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.953333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.953391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.953409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.953440 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:22 crc kubenswrapper[4754]: I1011 03:07:22.953459 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:22Z","lastTransitionTime":"2025-10-11T03:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.057262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.057338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.057359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.057388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.057409 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.083329 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:23 crc kubenswrapper[4754]: E1011 03:07:23.083537 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.106069 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmk28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecb2bdbe-2019-41c4-ada8-c5aa0d32dd68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a2f31e219481f9cf7f854b5a6226c30f5dc4bb9ff9b9a7285114de5e03e726e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xns6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmk28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.132559 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a8e02c-87f2-4b76-9048-bbbececf3016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07fee93b686f680e78f2c392a8ea95fb28a171a0130269f2484d81f3a49a0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e87cccd0ee39f0f1e6b487b9ae7c2554ae450344d446258f1b11e698817e54db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491b2d7e0cdf5c43ad012f43392250898e46db10bc4735e105361ba39636d8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa4f065ba8779d5bac794250c2319ea9462c47ae5f2af5059125dea1d8773fe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.156786 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5910057-0a3d-4b16-bb7f-47f9a0ebbf43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bf617ea6a1f39c8d6551a19ec4392b52060a96602a83540206be9cb75e2c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a498d02569ce55ffdfdf427b9746b87dfd24829d110c348e37ba04e3d164b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0346b1337150861b64748742df74b1f3d1e8a1baea0854f0f367073afb2d118c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16a98cd0e4688f362d0b3716c3b2fbc4019348fb7a05c5a10d3f6ddd046a270\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3deb346048930af30a44cb748db90998f2d4347e43257d0cf7b51f0b73572fc4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1011 03:06:11.997811 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1011 03:06:11.997835 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3539751986/tls.crt::/tmp/serving-cert-3539751986/tls.key\\\\\\\"\\\\nI1011 03:06:11.997986 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1011 03:06:11.999079 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999132 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:06:11.999173 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999187 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:06:11.999212 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1011 03:06:11.999218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1011 03:06:11.999318 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:06:11.999330 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1011 03:06:12.015618 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1011 03:06:12.015902 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.015893 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1011 03:06:12.016556 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21681166d8158bd40d61803d17e9977f80edc40524f267719073fa1d44dfafc3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc445441dbc81f3cda2dd783061e5ed160bb58c5559ef8c0f0157b88a25297a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.161749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.161796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.161809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.161833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.161849 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.176645 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.199591 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-r59xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef19b293-a4d7-44f5-a26d-4daf0e558d6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8aec4f41b588c25ebd0fdcc0eb0d82a25c7c62d049ff143e72c4e9c7d81aa8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b710fd72ad45e8057df919bca3d73c75b04abd16b1457bcd8c432fc599cce904\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c58b8e07663bf0eb8e3841dee08dc1601bc2daf15134c8aab93c12f27605ca6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da661317df5130de7c27b6bad29772891dabe352579bc47474cb2c9efec3428c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab00354d215aebb58d1f3300e130775d6a4f3ad82e263ee27ff4d8d37af77f3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://128b961d951b300fb6c29356f3728a9978ca6dc6e29c041525058004311e9a55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346b125a21eb95b3d9bfe579627642c7beb0f51a7bb5b3c5cae2b30ca12e0eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9bs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-r59xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.216572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6004db5-3e79-4c17-87b3-a38f1ead335d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0447cd338efa23a1f3eb00400a08348eb26779f81729d5ca4593edba3cefdf38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b92231bd8dd440d51c345f92e937da7fd6b9dad5a8fca65d1b22c4e8d7856857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.235870 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b220c2d-b3e5-4a77-9fb7-8dac2d16cecc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60400b8ee9c000f4485a33d51cabbb3c5c63c17eb6b876a8d3dbdd290cccd447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a78caca201a3418712e60e9f96d19df06dc45701baa6369c48f8fd43109125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://280dc1d160bbd621d26f7b930cc1f3d763d47f1e9629e8ccdd7d33ec23ed5847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81a1b91eecca1379ff0065b165b11f0dce42b31762825db7ca9a74d01151791e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.258313 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.265850 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.265921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.265942 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.266012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.266037 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.279599 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b4183f54b3649c72617c5ed34e091f01b631f394b81c8de06c7cfe4e4670cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.305944 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:08Z\\\",\\\"message\\\":\\\"x2sf\\\\nI1011 03:07:08.034497 6762 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1011 03:07:08.034213 6762 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1011 03:07:08.034576 6762 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034723 6762 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-2nlvv in node crc\\\\nI1011 03:07:08.034729 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-2nlvv after 0 failed attempt(s)\\\\nI1011 03:07:08.034734 6762 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-2nlvv\\\\nI1011 03:07:08.034484 6762 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1011 03:07:08.034749 6762 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nF1011 03:07:08.034752 6762 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fai\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4t5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-42sbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.326781 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2nlvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd26d5b-a1b6-41fa-b285-d650e389d8d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:07:05Z\\\",\\\"message\\\":\\\"2025-10-11T03:06:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46\\\\n2025-10-11T03:06:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_251154e7-6c0f-4070-9ee8-6d32ca1c4e46 to /host/opt/cni/bin/\\\\n2025-10-11T03:06:20Z [verbose] multus-daemon started\\\\n2025-10-11T03:06:20Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:07:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6f5mb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2nlvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.345998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93810265-58d5-4b44-a368-8b78981b9f2f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0db0ea12e2d23195b467384f2feff0edd8849cd7b40fc8ba226dd997a0d4566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea54b9ddfcebbef9792811435b556ad8de3826b5456a95b2e5f7dcf209d56f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkrbv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nrg7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.369844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.369934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.369953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.370036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.370084 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.375719 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecff5b6d-7486-43d2-89fb-ff8ac1b89692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115a352d4ccc6792bab6e105ccedecffe743a3633396c02861bbdc203aa16ca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f08f28ed3c66581e51614a50a296f0350974ad537192cc8df57c34aa66f612d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a3306df389d2f15e0422a2a450137aa760a09af444472c5ce1d6506ae1ba765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59755c9c2365ce9544dae34e7403b8bf8f98016632671cb0045a506241cfb589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c39f81e0638e0f042fe8720677a028faacf0aef008e522a97efcc0e99c28e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee74d026ecc345191ac62762ec7148d0bfa23cc1822f36df133367323b31c4ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8a8a8535cb67713d5fb8e95827981e81fe5efcf493f1c2833c1d2a1911f980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79de7121b5427a2b3fbbbae3c91ab7fb5bbee4c6e596f3611f2108054b8338f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:05:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:05:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.400880 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.426151 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8db2a74454bbe34157bed8358a6b41ce427846c13eb1d80acb4c63c8c0f7e2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0acbbff163aac717dd2a5d1b6fad229d5d55a7a597f9d378ec5233b6c97ea4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.450291 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef866bf9-b2ed-4ab9-922b-68f7373b6c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69222f07d2acbd3d8318ce784867a97f550737d2ed147541640c7d2e5b1a6566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-px2sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.468140 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nxj2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb406ab-5a49-485b-a5a5-af1d48675ae8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e92241c278adab291e4830ef675f2b566ce58a8eba56019514792b9c8abe25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff75d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nxj2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.474547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.474614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.474627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.474652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.474666 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.488216 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6s82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5e7151-ac6f-4da4-9cc8-39639136e005\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6s82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.510037 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3411a5b9e78bf38cf9782b693696380167d234e8fe0d27e17f0b90bed5c8dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:07:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.579255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.579334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.579361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.579396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.579418 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.682788 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.682860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.682879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.682904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.682920 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.785652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.785719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.785739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.785800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.785824 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.889013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.889101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.889132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.889175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.889198 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.992959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.993074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.993092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.993121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:23 crc kubenswrapper[4754]: I1011 03:07:23.993142 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:23Z","lastTransitionTime":"2025-10-11T03:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.083003 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.083148 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.083253 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:24 crc kubenswrapper[4754]: E1011 03:07:24.083372 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:24 crc kubenswrapper[4754]: E1011 03:07:24.083548 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:24 crc kubenswrapper[4754]: E1011 03:07:24.083658 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.097339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.097396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.097416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.097449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.097470 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.201357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.201419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.201444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.201472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.201491 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.304658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.304713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.304725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.304745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.304759 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.408774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.408836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.408856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.408882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.408900 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.513174 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.513241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.513269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.513307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.513332 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.617086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.617154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.617173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.617201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.617223 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.720554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.720629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.720652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.720680 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.720699 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.824251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.824300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.824319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.824341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.824356 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.927255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.927341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.927365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.927398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:24 crc kubenswrapper[4754]: I1011 03:07:24.927420 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:24Z","lastTransitionTime":"2025-10-11T03:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.031599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.031642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.031659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.031681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.031695 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.083129 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:25 crc kubenswrapper[4754]: E1011 03:07:25.083489 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.134700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.134777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.134792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.134816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.134833 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.237918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.237999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.238008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.238029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.238040 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.341696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.341779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.341800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.341835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.341854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.444078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.444126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.444135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.444151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.444162 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.550027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.550078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.550091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.550114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.550129 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.653480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.653531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.653550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.653575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.653596 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.758412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.758473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.758494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.758523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.758544 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.862392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.862474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.862494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.862526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.862547 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.966128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.966207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.966235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.966269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:25 crc kubenswrapper[4754]: I1011 03:07:25.966291 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:25Z","lastTransitionTime":"2025-10-11T03:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.069717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.069786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.069804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.069833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.069854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.083228 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.083354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.083221 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:26 crc kubenswrapper[4754]: E1011 03:07:26.083412 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:26 crc kubenswrapper[4754]: E1011 03:07:26.083605 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:26 crc kubenswrapper[4754]: E1011 03:07:26.083707 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.174299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.174382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.174406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.174432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.174452 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.277889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.277960 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.278029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.278054 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.278073 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.381866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.381947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.382008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.382048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.382073 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.485257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.485319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.485341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.485369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.485389 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.588176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.588247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.588267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.588292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.588315 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.691169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.691233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.691251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.691277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.691297 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.795851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.795934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.795957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.796073 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.796101 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.899703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.899773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.899796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.899826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:26 crc kubenswrapper[4754]: I1011 03:07:26.899844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:26Z","lastTransitionTime":"2025-10-11T03:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.003849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.003949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.004006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.004034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.004052 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.083553 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:27 crc kubenswrapper[4754]: E1011 03:07:27.083850 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.106785 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.106856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.106876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.106903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.106926 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.209897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.209949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.210001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.210025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.210088 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.313134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.313218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.313237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.313267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.313285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.416893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.417022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.417052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.417085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.417110 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.521170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.521232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.521251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.521279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.521296 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.658016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.658090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.658123 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.658152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.658172 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.762129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.762194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.762211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.762232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.762248 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.865843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.865902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.865920 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.865949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.866016 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.976900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.977018 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.977057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.977222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:27 crc kubenswrapper[4754]: I1011 03:07:27.977274 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:27Z","lastTransitionTime":"2025-10-11T03:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.081156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.081243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.081266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.081298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.081318 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.083517 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.083631 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:28 crc kubenswrapper[4754]: E1011 03:07:28.083667 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:28 crc kubenswrapper[4754]: E1011 03:07:28.084062 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.084093 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:28 crc kubenswrapper[4754]: E1011 03:07:28.084672 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.183938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.184024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.184043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.184068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.184086 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.287411 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.287448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.287458 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.287553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.287569 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.390327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.390372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.390388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.390412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.390429 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.492703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.492767 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.492786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.492814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.492832 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.596229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.596316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.596337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.596379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.596399 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.699681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.699719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.699728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.699742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.699753 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.802737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.802783 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.802795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.802812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.802825 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.905398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.905480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.905504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.905538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:28 crc kubenswrapper[4754]: I1011 03:07:28.905565 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:28Z","lastTransitionTime":"2025-10-11T03:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.008098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.008170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.008189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.008219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.008238 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.083363 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:29 crc kubenswrapper[4754]: E1011 03:07:29.083584 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.111544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.111614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.111633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.111660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.111684 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.214621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.214877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.214909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.214936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.214955 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.317925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.318000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.318018 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.318039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.318053 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.421436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.421496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.421512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.421535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.421553 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.524310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.524397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.524426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.524465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.524506 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.628011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.628107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.628133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.628157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.628175 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.731086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.731135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.731155 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.731178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.731195 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.834090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.834173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.834191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.834222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.834241 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.937552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.937655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.937673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.937699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:29 crc kubenswrapper[4754]: I1011 03:07:29.937723 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:29Z","lastTransitionTime":"2025-10-11T03:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.030126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.030202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.030224 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.030256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.030279 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:30Z","lastTransitionTime":"2025-10-11T03:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.064162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.064232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.064250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.064273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.064290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:07:30Z","lastTransitionTime":"2025-10-11T03:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.083174 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.083222 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:30 crc kubenswrapper[4754]: E1011 03:07:30.083377 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.083405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:30 crc kubenswrapper[4754]: E1011 03:07:30.083551 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:30 crc kubenswrapper[4754]: E1011 03:07:30.083708 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.105318 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp"] Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.106062 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.109738 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.109936 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.110173 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.114945 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.142651 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.142617827 podStartE2EDuration="1m18.142617827s" podCreationTimestamp="2025-10-11 03:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.140188489 +0000 UTC m=+97.699133274" watchObservedRunningTime="2025-10-11 03:07:30.142617827 +0000 UTC m=+97.701562662" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.191353 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-r59xj" podStartSLOduration=73.191312454 podStartE2EDuration="1m13.191312454s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.191222671 +0000 UTC m=+97.750167476" watchObservedRunningTime="2025-10-11 03:07:30.191312454 +0000 UTC m=+97.750257249" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.217204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/054c3f1b-347e-4ed3-a2f5-02192089826b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.217395 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054c3f1b-347e-4ed3-a2f5-02192089826b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.217518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.217583 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/054c3f1b-347e-4ed3-a2f5-02192089826b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.217661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.226572 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.226547383 podStartE2EDuration="1m17.226547383s" podCreationTimestamp="2025-10-11 03:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.225212955 +0000 UTC m=+97.784157760" watchObservedRunningTime="2025-10-11 03:07:30.226547383 +0000 UTC m=+97.785492178" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.226749 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.226742508 podStartE2EDuration="23.226742508s" podCreationTimestamp="2025-10-11 03:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.209905416 +0000 UTC m=+97.768850261" watchObservedRunningTime="2025-10-11 03:07:30.226742508 +0000 UTC m=+97.785687303" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.302916 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2nlvv" podStartSLOduration=73.302888645 podStartE2EDuration="1m13.302888645s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.302586046 +0000 UTC m=+97.861530851" watchObservedRunningTime="2025-10-11 03:07:30.302888645 +0000 UTC m=+97.861833440" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.316803 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nrg7p" podStartSLOduration=73.316777984 podStartE2EDuration="1m13.316777984s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.316328252 +0000 UTC m=+97.875273067" watchObservedRunningTime="2025-10-11 03:07:30.316777984 +0000 UTC m=+97.875722769" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.318756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054c3f1b-347e-4ed3-a2f5-02192089826b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.318863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.318911 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/054c3f1b-347e-4ed3-a2f5-02192089826b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.318993 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.319023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.319038 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/054c3f1b-347e-4ed3-a2f5-02192089826b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.319098 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/054c3f1b-347e-4ed3-a2f5-02192089826b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.320067 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/054c3f1b-347e-4ed3-a2f5-02192089826b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.332917 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/054c3f1b-347e-4ed3-a2f5-02192089826b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.340487 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/054c3f1b-347e-4ed3-a2f5-02192089826b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jr4lp\" (UID: \"054c3f1b-347e-4ed3-a2f5-02192089826b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.347363 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.347352203 podStartE2EDuration="1m16.347352203s" podCreationTimestamp="2025-10-11 03:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.346980942 +0000 UTC m=+97.905925727" watchObservedRunningTime="2025-10-11 03:07:30.347352203 +0000 UTC m=+97.906296988" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.393042 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podStartSLOduration=73.393018965 podStartE2EDuration="1m13.393018965s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.392172511 +0000 UTC m=+97.951117296" watchObservedRunningTime="2025-10-11 03:07:30.393018965 +0000 UTC m=+97.951963750" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.422293 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-nxj2n" podStartSLOduration=73.422270686 podStartE2EDuration="1m13.422270686s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.403432947 +0000 UTC m=+97.962377732" watchObservedRunningTime="2025-10-11 03:07:30.422270686 +0000 UTC m=+97.981215471" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.431460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.453059 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-pmk28" podStartSLOduration=73.453037619 podStartE2EDuration="1m13.453037619s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.452934506 +0000 UTC m=+98.011879291" watchObservedRunningTime="2025-10-11 03:07:30.453037619 +0000 UTC m=+98.011982414" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.489354 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.489332198 podStartE2EDuration="41.489332198s" podCreationTimestamp="2025-10-11 03:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.489300147 +0000 UTC m=+98.048244942" watchObservedRunningTime="2025-10-11 03:07:30.489332198 +0000 UTC m=+98.048276983" Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.674229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" event={"ID":"054c3f1b-347e-4ed3-a2f5-02192089826b","Type":"ContainerStarted","Data":"6299c1faddc9bdbc7638c97d6a3728ae1b6fc54c92c30c5ebdfca24af3b8f43c"} Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.674281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" event={"ID":"054c3f1b-347e-4ed3-a2f5-02192089826b","Type":"ContainerStarted","Data":"36e5c619a366f9e20445a13de0bbdf6af6dc82b28174a76daa7b9763ec9ec9b6"} Oct 11 03:07:30 crc kubenswrapper[4754]: I1011 03:07:30.692927 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jr4lp" podStartSLOduration=73.692908532 podStartE2EDuration="1m13.692908532s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:30.692037378 +0000 UTC m=+98.250982163" watchObservedRunningTime="2025-10-11 03:07:30.692908532 +0000 UTC m=+98.251853327" Oct 11 03:07:31 crc kubenswrapper[4754]: I1011 03:07:31.083338 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:31 crc kubenswrapper[4754]: E1011 03:07:31.083573 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:32 crc kubenswrapper[4754]: I1011 03:07:32.083156 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:32 crc kubenswrapper[4754]: I1011 03:07:32.083237 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:32 crc kubenswrapper[4754]: I1011 03:07:32.083311 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:32 crc kubenswrapper[4754]: E1011 03:07:32.083334 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:32 crc kubenswrapper[4754]: E1011 03:07:32.083473 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:32 crc kubenswrapper[4754]: E1011 03:07:32.083593 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:33 crc kubenswrapper[4754]: I1011 03:07:33.083566 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:33 crc kubenswrapper[4754]: E1011 03:07:33.086063 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:33 crc kubenswrapper[4754]: I1011 03:07:33.086615 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:33 crc kubenswrapper[4754]: E1011 03:07:33.086888 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:07:34 crc kubenswrapper[4754]: I1011 03:07:34.083511 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:34 crc kubenswrapper[4754]: E1011 03:07:34.083618 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:34 crc kubenswrapper[4754]: I1011 03:07:34.083679 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:34 crc kubenswrapper[4754]: E1011 03:07:34.083732 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:34 crc kubenswrapper[4754]: I1011 03:07:34.084039 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:34 crc kubenswrapper[4754]: E1011 03:07:34.084220 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:35 crc kubenswrapper[4754]: I1011 03:07:35.083910 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:35 crc kubenswrapper[4754]: E1011 03:07:35.084237 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:35 crc kubenswrapper[4754]: I1011 03:07:35.885726 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:35 crc kubenswrapper[4754]: E1011 03:07:35.886092 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:07:35 crc kubenswrapper[4754]: E1011 03:07:35.886254 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs podName:fb5e7151-ac6f-4da4-9cc8-39639136e005 nodeName:}" failed. No retries permitted until 2025-10-11 03:08:39.886210053 +0000 UTC m=+167.445155028 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs") pod "network-metrics-daemon-b6s82" (UID: "fb5e7151-ac6f-4da4-9cc8-39639136e005") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:07:36 crc kubenswrapper[4754]: I1011 03:07:36.083185 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:36 crc kubenswrapper[4754]: I1011 03:07:36.083294 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:36 crc kubenswrapper[4754]: I1011 03:07:36.083328 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:36 crc kubenswrapper[4754]: E1011 03:07:36.083451 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:36 crc kubenswrapper[4754]: E1011 03:07:36.083581 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:36 crc kubenswrapper[4754]: E1011 03:07:36.083763 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:37 crc kubenswrapper[4754]: I1011 03:07:37.083243 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:37 crc kubenswrapper[4754]: E1011 03:07:37.083435 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:38 crc kubenswrapper[4754]: I1011 03:07:38.083731 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:38 crc kubenswrapper[4754]: I1011 03:07:38.083841 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:38 crc kubenswrapper[4754]: E1011 03:07:38.083986 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:38 crc kubenswrapper[4754]: I1011 03:07:38.083869 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:38 crc kubenswrapper[4754]: E1011 03:07:38.084146 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:38 crc kubenswrapper[4754]: E1011 03:07:38.084287 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:39 crc kubenswrapper[4754]: I1011 03:07:39.083278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:39 crc kubenswrapper[4754]: E1011 03:07:39.083545 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:40 crc kubenswrapper[4754]: I1011 03:07:40.083447 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:40 crc kubenswrapper[4754]: I1011 03:07:40.083498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:40 crc kubenswrapper[4754]: I1011 03:07:40.083526 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:40 crc kubenswrapper[4754]: E1011 03:07:40.083590 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:40 crc kubenswrapper[4754]: E1011 03:07:40.083798 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:40 crc kubenswrapper[4754]: E1011 03:07:40.083873 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:41 crc kubenswrapper[4754]: I1011 03:07:41.083753 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:41 crc kubenswrapper[4754]: E1011 03:07:41.084184 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:42 crc kubenswrapper[4754]: I1011 03:07:42.083445 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:42 crc kubenswrapper[4754]: I1011 03:07:42.083527 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:42 crc kubenswrapper[4754]: I1011 03:07:42.083552 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:42 crc kubenswrapper[4754]: E1011 03:07:42.083694 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:42 crc kubenswrapper[4754]: E1011 03:07:42.083832 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:42 crc kubenswrapper[4754]: E1011 03:07:42.084253 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:43 crc kubenswrapper[4754]: I1011 03:07:43.083003 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:43 crc kubenswrapper[4754]: E1011 03:07:43.083932 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:44 crc kubenswrapper[4754]: I1011 03:07:44.083488 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:44 crc kubenswrapper[4754]: I1011 03:07:44.083591 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:44 crc kubenswrapper[4754]: I1011 03:07:44.083618 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:44 crc kubenswrapper[4754]: E1011 03:07:44.083651 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:44 crc kubenswrapper[4754]: E1011 03:07:44.083726 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:44 crc kubenswrapper[4754]: E1011 03:07:44.083866 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:45 crc kubenswrapper[4754]: I1011 03:07:45.082792 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:45 crc kubenswrapper[4754]: E1011 03:07:45.083092 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:45 crc kubenswrapper[4754]: I1011 03:07:45.084920 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:45 crc kubenswrapper[4754]: E1011 03:07:45.085315 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-42sbf_openshift-ovn-kubernetes(0dea8260-bc1e-4ad6-96d3-f9d202b125c7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" Oct 11 03:07:46 crc kubenswrapper[4754]: I1011 03:07:46.083524 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:46 crc kubenswrapper[4754]: I1011 03:07:46.083525 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:46 crc kubenswrapper[4754]: I1011 03:07:46.083673 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:46 crc kubenswrapper[4754]: E1011 03:07:46.083701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:46 crc kubenswrapper[4754]: E1011 03:07:46.083890 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:46 crc kubenswrapper[4754]: E1011 03:07:46.084053 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:47 crc kubenswrapper[4754]: I1011 03:07:47.083335 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:47 crc kubenswrapper[4754]: E1011 03:07:47.083518 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:48 crc kubenswrapper[4754]: I1011 03:07:48.083566 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:48 crc kubenswrapper[4754]: I1011 03:07:48.083621 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:48 crc kubenswrapper[4754]: I1011 03:07:48.083707 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:48 crc kubenswrapper[4754]: E1011 03:07:48.084394 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:48 crc kubenswrapper[4754]: E1011 03:07:48.084527 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:48 crc kubenswrapper[4754]: E1011 03:07:48.084619 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:49 crc kubenswrapper[4754]: I1011 03:07:49.083157 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:49 crc kubenswrapper[4754]: E1011 03:07:49.083375 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:50 crc kubenswrapper[4754]: I1011 03:07:50.083135 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:50 crc kubenswrapper[4754]: I1011 03:07:50.083214 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:50 crc kubenswrapper[4754]: E1011 03:07:50.083269 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:50 crc kubenswrapper[4754]: I1011 03:07:50.083135 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:50 crc kubenswrapper[4754]: E1011 03:07:50.083341 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:50 crc kubenswrapper[4754]: E1011 03:07:50.083584 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:51 crc kubenswrapper[4754]: I1011 03:07:51.083846 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:51 crc kubenswrapper[4754]: E1011 03:07:51.084167 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.083082 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.083134 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.083105 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:52 crc kubenswrapper[4754]: E1011 03:07:52.083268 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:52 crc kubenswrapper[4754]: E1011 03:07:52.083348 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:52 crc kubenswrapper[4754]: E1011 03:07:52.083511 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.759854 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/1.log" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.760702 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/0.log" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.760776 4754 generic.go:334] "Generic (PLEG): container finished" podID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" containerID="5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2" exitCode=1 Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.760826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerDied","Data":"5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2"} Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.760883 4754 scope.go:117] "RemoveContainer" containerID="acec3d553f0e00cdeeab78975817203fd574cb0d69b80715a524229233eeac5b" Oct 11 03:07:52 crc kubenswrapper[4754]: I1011 03:07:52.761602 4754 scope.go:117] "RemoveContainer" containerID="5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2" Oct 11 03:07:52 crc kubenswrapper[4754]: E1011 03:07:52.761893 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2nlvv_openshift-multus(3cd26d5b-a1b6-41fa-b285-d650e389d8d4)\"" pod="openshift-multus/multus-2nlvv" podUID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" Oct 11 03:07:53 crc kubenswrapper[4754]: E1011 03:07:53.031171 4754 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 11 03:07:53 crc kubenswrapper[4754]: I1011 03:07:53.083655 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:53 crc kubenswrapper[4754]: E1011 03:07:53.086427 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:53 crc kubenswrapper[4754]: E1011 03:07:53.174895 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:07:53 crc kubenswrapper[4754]: I1011 03:07:53.765173 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/1.log" Oct 11 03:07:54 crc kubenswrapper[4754]: I1011 03:07:54.083239 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:54 crc kubenswrapper[4754]: I1011 03:07:54.083289 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:54 crc kubenswrapper[4754]: E1011 03:07:54.083404 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:54 crc kubenswrapper[4754]: I1011 03:07:54.083298 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:54 crc kubenswrapper[4754]: E1011 03:07:54.083529 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:54 crc kubenswrapper[4754]: E1011 03:07:54.083796 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:55 crc kubenswrapper[4754]: I1011 03:07:55.083665 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:55 crc kubenswrapper[4754]: E1011 03:07:55.083832 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.083270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.083319 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.083272 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:56 crc kubenswrapper[4754]: E1011 03:07:56.083440 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:56 crc kubenswrapper[4754]: E1011 03:07:56.083544 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:56 crc kubenswrapper[4754]: E1011 03:07:56.084045 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.085257 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.780619 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/3.log" Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.785714 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerStarted","Data":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:07:56 crc kubenswrapper[4754]: I1011 03:07:56.786326 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:07:57 crc kubenswrapper[4754]: I1011 03:07:57.083362 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:57 crc kubenswrapper[4754]: E1011 03:07:57.083480 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:57 crc kubenswrapper[4754]: I1011 03:07:57.194988 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podStartSLOduration=100.19493617 podStartE2EDuration="1m40.19493617s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:07:56.836239012 +0000 UTC m=+124.395183797" watchObservedRunningTime="2025-10-11 03:07:57.19493617 +0000 UTC m=+124.753880975" Oct 11 03:07:57 crc kubenswrapper[4754]: I1011 03:07:57.195504 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b6s82"] Oct 11 03:07:57 crc kubenswrapper[4754]: I1011 03:07:57.789033 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:57 crc kubenswrapper[4754]: E1011 03:07:57.789781 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:07:58 crc kubenswrapper[4754]: I1011 03:07:58.082677 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:07:58 crc kubenswrapper[4754]: I1011 03:07:58.082735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:07:58 crc kubenswrapper[4754]: I1011 03:07:58.082694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:07:58 crc kubenswrapper[4754]: E1011 03:07:58.082814 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:07:58 crc kubenswrapper[4754]: E1011 03:07:58.083010 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:07:58 crc kubenswrapper[4754]: E1011 03:07:58.083053 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:07:58 crc kubenswrapper[4754]: E1011 03:07:58.176528 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:07:59 crc kubenswrapper[4754]: I1011 03:07:59.082672 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:07:59 crc kubenswrapper[4754]: E1011 03:07:59.082849 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:08:00 crc kubenswrapper[4754]: I1011 03:08:00.082797 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:00 crc kubenswrapper[4754]: E1011 03:08:00.082998 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:08:00 crc kubenswrapper[4754]: I1011 03:08:00.082822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:00 crc kubenswrapper[4754]: E1011 03:08:00.083089 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:08:00 crc kubenswrapper[4754]: I1011 03:08:00.082797 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:00 crc kubenswrapper[4754]: E1011 03:08:00.083162 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:08:01 crc kubenswrapper[4754]: I1011 03:08:01.083331 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:01 crc kubenswrapper[4754]: E1011 03:08:01.083556 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:08:02 crc kubenswrapper[4754]: I1011 03:08:02.083207 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:02 crc kubenswrapper[4754]: I1011 03:08:02.083260 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:02 crc kubenswrapper[4754]: I1011 03:08:02.083292 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:02 crc kubenswrapper[4754]: E1011 03:08:02.083371 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:08:02 crc kubenswrapper[4754]: E1011 03:08:02.083637 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:08:02 crc kubenswrapper[4754]: E1011 03:08:02.083592 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:08:03 crc kubenswrapper[4754]: I1011 03:08:03.083667 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:03 crc kubenswrapper[4754]: E1011 03:08:03.087071 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:08:03 crc kubenswrapper[4754]: I1011 03:08:03.087625 4754 scope.go:117] "RemoveContainer" containerID="5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2" Oct 11 03:08:03 crc kubenswrapper[4754]: E1011 03:08:03.177525 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:08:03 crc kubenswrapper[4754]: I1011 03:08:03.812003 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/1.log" Oct 11 03:08:03 crc kubenswrapper[4754]: I1011 03:08:03.812544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerStarted","Data":"c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf"} Oct 11 03:08:04 crc kubenswrapper[4754]: I1011 03:08:04.083034 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:04 crc kubenswrapper[4754]: I1011 03:08:04.083144 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:04 crc kubenswrapper[4754]: E1011 03:08:04.083183 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:08:04 crc kubenswrapper[4754]: I1011 03:08:04.083045 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:04 crc kubenswrapper[4754]: E1011 03:08:04.083283 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:08:04 crc kubenswrapper[4754]: E1011 03:08:04.083480 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:08:05 crc kubenswrapper[4754]: I1011 03:08:05.083061 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:05 crc kubenswrapper[4754]: E1011 03:08:05.083352 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:08:06 crc kubenswrapper[4754]: I1011 03:08:06.083662 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:06 crc kubenswrapper[4754]: E1011 03:08:06.083885 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:08:06 crc kubenswrapper[4754]: I1011 03:08:06.084280 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:06 crc kubenswrapper[4754]: E1011 03:08:06.084385 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:08:06 crc kubenswrapper[4754]: I1011 03:08:06.084663 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:06 crc kubenswrapper[4754]: E1011 03:08:06.084796 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:08:07 crc kubenswrapper[4754]: I1011 03:08:07.083472 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:07 crc kubenswrapper[4754]: E1011 03:08:07.083687 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6s82" podUID="fb5e7151-ac6f-4da4-9cc8-39639136e005" Oct 11 03:08:08 crc kubenswrapper[4754]: I1011 03:08:08.083082 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:08 crc kubenswrapper[4754]: I1011 03:08:08.083191 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:08 crc kubenswrapper[4754]: I1011 03:08:08.083115 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:08 crc kubenswrapper[4754]: E1011 03:08:08.083328 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:08:08 crc kubenswrapper[4754]: E1011 03:08:08.083490 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:08:08 crc kubenswrapper[4754]: E1011 03:08:08.083605 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:08:09 crc kubenswrapper[4754]: I1011 03:08:09.083270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:09 crc kubenswrapper[4754]: I1011 03:08:09.086474 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 11 03:08:09 crc kubenswrapper[4754]: I1011 03:08:09.087229 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.083176 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.083254 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.083176 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.088726 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.089365 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.094621 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.094627 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.614586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.678574 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.679871 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.683033 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.683794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.686605 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dhcrp"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.687412 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.688377 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.689159 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.691749 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.692482 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.700462 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-42nc2"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.701412 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.702077 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.703652 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.706282 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.708807 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.709066 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.709169 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.709333 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.715912 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.719336 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.719905 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.720251 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.720479 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.720794 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721071 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721185 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721215 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721442 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721612 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721702 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.721766 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.736629 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.737007 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.755533 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.756026 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.765836 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.766408 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.770705 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.774220 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.782120 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.782238 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.782464 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.782559 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.782718 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.784798 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785118 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785137 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785393 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785462 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785493 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785640 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785680 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785789 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785812 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785880 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785918 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.785978 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.786320 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.786443 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.786561 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.786976 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.792334 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.793350 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.799199 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-57clb"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.802917 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.819038 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.819516 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-h69hd"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.819929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.820382 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.820640 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.820709 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.821742 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.822168 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.822494 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.822580 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.822695 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.823048 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.823406 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.823412 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.824212 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.824539 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6bbxp"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.824824 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.825083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.825269 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4qb6"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.825620 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.831950 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.832783 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.832945 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bppnf"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833284 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833490 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833578 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833590 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833617 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833756 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dhcrp"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833920 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.833986 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834030 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834087 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834164 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834187 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834206 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834313 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834325 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834442 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834556 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834593 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834627 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.834695 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.835326 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.836334 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.836485 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.836605 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.836686 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.844241 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.844853 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.844974 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845058 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845130 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845181 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845229 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845353 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845556 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845700 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845891 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.846116 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.845895 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.844997 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.846914 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.854096 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.854700 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.855021 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.855068 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.855030 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.855307 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.856056 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.856128 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.856233 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.859918 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.860995 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.861604 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.861604 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.862180 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.862804 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.888422 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.889056 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.891730 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.895069 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.896273 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vcpwr"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.898784 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902112 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902148 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3970eb8-365e-4236-95b0-d556e52eb675-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-trusted-ca-bundle\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902205 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a591a71f-4786-44ca-ba67-f289f2118e79-config\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.902229 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.918667 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.918712 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.923786 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.924015 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.924370 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.924757 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.925777 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-464tm\" (UniqueName: \"kubernetes.io/projected/aeac3b86-721d-47c3-99a4-523e11d3b89b-kube-api-access-464tm\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.925847 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.925876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.925898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926034 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926075 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-encryption-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926143 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-trusted-ca\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq69t\" (UniqueName: \"kubernetes.io/projected/f6e20429-5a1a-469e-9826-3d5561f8c627-kube-api-access-wq69t\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-trusted-ca\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-config\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926345 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a591a71f-4786-44ca-ba67-f289f2118e79-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926365 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-serving-cert\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926408 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-client\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926431 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpx8k\" (UniqueName: \"kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926452 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b67t\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-kube-api-access-9b67t\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926510 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926553 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gszkg\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-kube-api-access-gszkg\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926665 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a591a71f-4786-44ca-ba67-f289f2118e79-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926691 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-dir\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t42ql\" (UniqueName: \"kubernetes.io/projected/2243842d-42a7-446c-a5a9-7d57d6fce732-kube-api-access-t42ql\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926734 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e720967-ded2-443d-a9ae-dadbeebebdaa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926805 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htngs\" (UniqueName: \"kubernetes.io/projected/6fe7f8e1-0b7e-44df-820a-924402b9b889-kube-api-access-htngs\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926822 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-service-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926846 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-metrics-tls\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7xm8\" (UniqueName: \"kubernetes.io/projected/f9ca86b2-12d2-4285-a892-6b85beb92b72-kube-api-access-t7xm8\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-client\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926950 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/424e4796-34b6-4b66-94eb-a3c2918667de-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.926989 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-audit-dir\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927008 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0cc42e-e539-4597-a986-0cdf90611c82-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927040 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqkcz\" (UniqueName: \"kubernetes.io/projected/8e720967-ded2-443d-a9ae-dadbeebebdaa-kube-api-access-vqkcz\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927128 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpj7x\" (UniqueName: \"kubernetes.io/projected/9b8d25f2-185a-454a-9601-10f51d68412b-kube-api-access-gpj7x\") pod \"downloads-7954f5f757-h69hd\" (UID: \"9b8d25f2-185a-454a-9601-10f51d68412b\") " pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927171 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927239 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927265 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927336 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927345 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrl7s\" (UniqueName: \"kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-encryption-config\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927512 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-client\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927538 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-serving-cert\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927563 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fe7f8e1-0b7e-44df-820a-924402b9b889-serving-cert\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927588 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927641 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-audit\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927673 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e20429-5a1a-469e-9826-3d5561f8c627-serving-cert\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d8e642c-404d-4b20-8443-7e00e75f3c1c-machine-approver-tls\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927755 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gks\" (UniqueName: \"kubernetes.io/projected/5d8e642c-404d-4b20-8443-7e00e75f3c1c-kube-api-access-c8gks\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927777 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927821 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-config\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927853 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hh72\" (UniqueName: \"kubernetes.io/projected/7acf8241-6d60-494a-847d-ffab61ae4242-kube-api-access-8hh72\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927888 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qf2l\" (UniqueName: \"kubernetes.io/projected/b3970eb8-365e-4236-95b0-d556e52eb675-kube-api-access-7qf2l\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927937 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hnzg\" (UniqueName: \"kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.927975 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928046 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7acf8241-6d60-494a-847d-ffab61ae4242-serving-cert\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3970eb8-365e-4236-95b0-d556e52eb675-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928101 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-serving-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-config\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928149 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-serving-cert\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-images\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928287 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928301 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-policies\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928360 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928376 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2krw\" (UniqueName: \"kubernetes.io/projected/ce0cc42e-e539-4597-a986-0cdf90611c82-kube-api-access-v2krw\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928417 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928437 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928505 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928552 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrxqq\" (UniqueName: \"kubernetes.io/projected/53ad7084-a584-4a86-8abd-d6419f9072c1-kube-api-access-hrxqq\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928572 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0cc42e-e539-4597-a986-0cdf90611c82-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928829 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928848 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-image-import-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928875 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-auth-proxy-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928895 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928916 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928934 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7acf8241-6d60-494a-847d-ffab61ae4242-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.928987 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-service-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53ad7084-a584-4a86-8abd-d6419f9072c1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prt8s\" (UniqueName: \"kubernetes.io/projected/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-kube-api-access-prt8s\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929045 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sr66\" (UniqueName: \"kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929083 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/424e4796-34b6-4b66-94eb-a3c2918667de-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929105 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-node-pullsecrets\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929122 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.929142 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.931086 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-config\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.931559 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-wq6lw"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.932433 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.932866 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.933552 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.933558 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.934268 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.936030 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.936838 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.938588 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.939464 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.940004 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.940582 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.943227 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xn7rq"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.943531 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.944060 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.944296 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.944557 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.946125 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.947849 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.948470 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.951619 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.952203 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.952292 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.952539 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6bbxp"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.953938 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-42nc2"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.954774 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.955380 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.955711 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-clhf8"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.956511 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.956895 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.958111 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vcpwr"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.958228 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.960540 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.962657 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.964957 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.966014 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.966595 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.967324 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.967891 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cdmps"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.969661 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.972882 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.973548 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.975378 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4qb6"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.980434 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bppnf"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.983329 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.985383 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.987478 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kl4ld"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.990034 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.990179 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.994240 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.996959 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dtnsb"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.998603 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4"] Oct 11 03:08:10 crc kubenswrapper[4754]: I1011 03:08:10.998614 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.001912 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-c5dg2"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.003167 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.003436 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cdmps"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.003577 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.004981 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.006148 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h69hd"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.008954 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.014114 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.016857 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.019181 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.020605 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.022185 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.022454 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.026045 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.027325 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xn7rq"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.028475 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dtnsb"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.029635 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.030623 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.031648 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032674 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032760 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-images\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-policies\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032890 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032923 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2krw\" (UniqueName: \"kubernetes.io/projected/ce0cc42e-e539-4597-a986-0cdf90611c82-kube-api-access-v2krw\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.032988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrxqq\" (UniqueName: \"kubernetes.io/projected/53ad7084-a584-4a86-8abd-d6419f9072c1-kube-api-access-hrxqq\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033154 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0cc42e-e539-4597-a986-0cdf90611c82-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033230 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-image-import-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033260 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-auth-proxy-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033292 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033331 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7acf8241-6d60-494a-847d-ffab61ae4242-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033395 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-service-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033455 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53ad7084-a584-4a86-8abd-d6419f9072c1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033480 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/424e4796-34b6-4b66-94eb-a3c2918667de-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033506 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prt8s\" (UniqueName: \"kubernetes.io/projected/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-kube-api-access-prt8s\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033525 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033548 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sr66\" (UniqueName: \"kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-node-pullsecrets\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033630 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-config\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033670 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3970eb8-365e-4236-95b0-d556e52eb675-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033745 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-trusted-ca-bundle\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033768 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a591a71f-4786-44ca-ba67-f289f2118e79-config\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-464tm\" (UniqueName: \"kubernetes.io/projected/aeac3b86-721d-47c3-99a4-523e11d3b89b-kube-api-access-464tm\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033807 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-encryption-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033827 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033869 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033893 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033912 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-trusted-ca\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033934 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-trusted-ca\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033957 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq69t\" (UniqueName: \"kubernetes.io/projected/f6e20429-5a1a-469e-9826-3d5561f8c627-kube-api-access-wq69t\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033985 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-policies\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033985 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.033994 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-config\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-images\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034079 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a591a71f-4786-44ca-ba67-f289f2118e79-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034104 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-serving-cert\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034108 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034131 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-client\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpx8k\" (UniqueName: \"kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b67t\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-kube-api-access-9b67t\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034275 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gszkg\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-kube-api-access-gszkg\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a591a71f-4786-44ca-ba67-f289f2118e79-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-dir\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t42ql\" (UniqueName: \"kubernetes.io/projected/2243842d-42a7-446c-a5a9-7d57d6fce732-kube-api-access-t42ql\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034414 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-service-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034438 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e720967-ded2-443d-a9ae-dadbeebebdaa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htngs\" (UniqueName: \"kubernetes.io/projected/6fe7f8e1-0b7e-44df-820a-924402b9b889-kube-api-access-htngs\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034477 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-metrics-tls\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034494 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-client\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034537 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034550 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7xm8\" (UniqueName: \"kubernetes.io/projected/f9ca86b2-12d2-4285-a892-6b85beb92b72-kube-api-access-t7xm8\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/424e4796-34b6-4b66-94eb-a3c2918667de-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034694 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqkcz\" (UniqueName: \"kubernetes.io/projected/8e720967-ded2-443d-a9ae-dadbeebebdaa-kube-api-access-vqkcz\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034725 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-audit-dir\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034750 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0cc42e-e539-4597-a986-0cdf90611c82-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034779 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpj7x\" (UniqueName: \"kubernetes.io/projected/9b8d25f2-185a-454a-9601-10f51d68412b-kube-api-access-gpj7x\") pod \"downloads-7954f5f757-h69hd\" (UID: \"9b8d25f2-185a-454a-9601-10f51d68412b\") " pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034803 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034832 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034857 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034883 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrl7s\" (UniqueName: \"kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.034912 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-encryption-config\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035009 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-client\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035036 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-serving-cert\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e720967-ded2-443d-a9ae-dadbeebebdaa-config\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fe7f8e1-0b7e-44df-820a-924402b9b889-serving-cert\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035143 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-audit\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035187 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e20429-5a1a-469e-9826-3d5561f8c627-serving-cert\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d8e642c-404d-4b20-8443-7e00e75f3c1c-machine-approver-tls\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035271 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gks\" (UniqueName: \"kubernetes.io/projected/5d8e642c-404d-4b20-8443-7e00e75f3c1c-kube-api-access-c8gks\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035315 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-config\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hh72\" (UniqueName: \"kubernetes.io/projected/7acf8241-6d60-494a-847d-ffab61ae4242-kube-api-access-8hh72\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035357 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035381 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qf2l\" (UniqueName: \"kubernetes.io/projected/b3970eb8-365e-4236-95b0-d556e52eb675-kube-api-access-7qf2l\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035403 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hnzg\" (UniqueName: \"kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035422 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035442 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7acf8241-6d60-494a-847d-ffab61ae4242-serving-cert\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035463 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3970eb8-365e-4236-95b0-d556e52eb675-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035487 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-serving-cert\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035507 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-serving-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-config\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.035543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.036383 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.036493 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.036823 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-service-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-config\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037871 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037906 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kl4ld"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037918 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-c5dg2"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.037983 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.038157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.038182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6e20429-5a1a-469e-9826-3d5561f8c627-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.038295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-node-pullsecrets\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.039545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.040062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-auth-proxy-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.040133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9ca86b2-12d2-4285-a892-6b85beb92b72-audit-dir\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.040473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-image-import-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.040743 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-clhf8"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.040773 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-cxhgj"] Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.041014 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.041714 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.041805 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-audit\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.042051 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.042813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2243842d-42a7-446c-a5a9-7d57d6fce732-audit-dir\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.043226 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.043172 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a591a71f-4786-44ca-ba67-f289f2118e79-config\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.043529 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.043537 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.043632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7acf8241-6d60-494a-847d-ffab61ae4242-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.044229 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fe7f8e1-0b7e-44df-820a-924402b9b889-serving-cert\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.044676 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-service-ca\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.044732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8e642c-404d-4b20-8443-7e00e75f3c1c-config\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.045602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-trusted-ca-bundle\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.046763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047078 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeac3b86-721d-47c3-99a4-523e11d3b89b-config\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047574 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047696 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047809 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3970eb8-365e-4236-95b0-d556e52eb675-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.047844 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048112 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-trusted-ca\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048243 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/424e4796-34b6-4b66-94eb-a3c2918667de-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048468 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/424e4796-34b6-4b66-94eb-a3c2918667de-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048465 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7acf8241-6d60-494a-847d-ffab61ae4242-serving-cert\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048673 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048712 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-etcd-client\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.048859 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049345 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-etcd-client\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-serving-ca\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049600 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-serving-cert\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049823 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d8e642c-404d-4b20-8443-7e00e75f3c1c-machine-approver-tls\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.049949 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe7f8e1-0b7e-44df-820a-924402b9b889-config\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.050000 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-encryption-config\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.050315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeac3b86-721d-47c3-99a4-523e11d3b89b-serving-cert\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.050480 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2243842d-42a7-446c-a5a9-7d57d6fce732-etcd-client\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.051100 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.051460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3970eb8-365e-4236-95b0-d556e52eb675-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.051595 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e720967-ded2-443d-a9ae-dadbeebebdaa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.052490 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.052537 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e20429-5a1a-469e-9826-3d5561f8c627-serving-cert\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.052562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53ad7084-a584-4a86-8abd-d6419f9072c1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.052687 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-encryption-config\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.052769 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.053332 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a591a71f-4786-44ca-ba67-f289f2118e79-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.053508 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.054266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9ca86b2-12d2-4285-a892-6b85beb92b72-serving-cert\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.054925 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.062612 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.091464 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.096152 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-trusted-ca\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.102437 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.122507 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.142395 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.161883 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.175205 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-metrics-tls\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.183530 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.202728 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.222558 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.242652 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.262258 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.272295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0cc42e-e539-4597-a986-0cdf90611c82-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.284411 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.301894 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.309929 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0cc42e-e539-4597-a986-0cdf90611c82-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.364115 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.384337 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.404100 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.422758 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.442520 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.462798 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.485023 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.502449 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.523477 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.543313 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.562862 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.582107 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.601991 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.623617 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.643029 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.664098 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.682407 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.703076 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.724271 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.742498 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.762606 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.783281 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.803443 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.824172 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.843542 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.863156 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.882921 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.903481 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.923363 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.943678 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.961146 4754 request.go:700] Waited for 1.016533805s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.963029 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 11 03:08:11 crc kubenswrapper[4754]: I1011 03:08:11.983071 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.003390 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.023882 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.043432 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.063738 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.083827 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.102258 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.122731 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.143605 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.164403 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.183123 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.203893 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.222910 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.244062 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.264030 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.284100 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.304566 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.323505 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.345077 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.364797 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.389361 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.403718 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.434190 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.443519 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.463999 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.483429 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.503898 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.524509 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.543687 4754 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.562414 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.582829 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.603192 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.622651 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.643001 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.662706 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.699572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2krw\" (UniqueName: \"kubernetes.io/projected/ce0cc42e-e539-4597-a986-0cdf90611c82-kube-api-access-v2krw\") pod \"kube-storage-version-migrator-operator-b67b599dd-ntpp8\" (UID: \"ce0cc42e-e539-4597-a986-0cdf90611c82\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.718017 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrxqq\" (UniqueName: \"kubernetes.io/projected/53ad7084-a584-4a86-8abd-d6419f9072c1-kube-api-access-hrxqq\") pod \"cluster-samples-operator-665b6dd947-zmxds\" (UID: \"53ad7084-a584-4a86-8abd-d6419f9072c1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.739347 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpx8k\" (UniqueName: \"kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k\") pod \"route-controller-manager-6576b87f9c-9cc52\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.743873 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.781187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prt8s\" (UniqueName: \"kubernetes.io/projected/55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5-kube-api-access-prt8s\") pod \"openshift-apiserver-operator-796bbdcf4f-cgb7f\" (UID: \"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.792099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b67t\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-kube-api-access-9b67t\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.798819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.809687 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.823748 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.828572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sr66\" (UniqueName: \"kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66\") pod \"oauth-openshift-558db77b4-v89kr\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.855177 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7xm8\" (UniqueName: \"kubernetes.io/projected/f9ca86b2-12d2-4285-a892-6b85beb92b72-kube-api-access-t7xm8\") pod \"apiserver-7bbb656c7d-q5dr7\" (UID: \"f9ca86b2-12d2-4285-a892-6b85beb92b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.864211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b0bf758-fa24-4d06-a7bc-52f8e1c05919-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fkmrd\" (UID: \"1b0bf758-fa24-4d06-a7bc-52f8e1c05919\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.881533 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.885279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpj7x\" (UniqueName: \"kubernetes.io/projected/9b8d25f2-185a-454a-9601-10f51d68412b-kube-api-access-gpj7x\") pod \"downloads-7954f5f757-h69hd\" (UID: \"9b8d25f2-185a-454a-9601-10f51d68412b\") " pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.907846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gszkg\" (UniqueName: \"kubernetes.io/projected/424e4796-34b6-4b66-94eb-a3c2918667de-kube-api-access-gszkg\") pod \"cluster-image-registry-operator-dc59b4c8b-qhjfs\" (UID: \"424e4796-34b6-4b66-94eb-a3c2918667de\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.930425 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a591a71f-4786-44ca-ba67-f289f2118e79-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mg6v2\" (UID: \"a591a71f-4786-44ca-ba67-f289f2118e79\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.956789 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.961097 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t42ql\" (UniqueName: \"kubernetes.io/projected/2243842d-42a7-446c-a5a9-7d57d6fce732-kube-api-access-t42ql\") pod \"apiserver-76f77b778f-42nc2\" (UID: \"2243842d-42a7-446c-a5a9-7d57d6fce732\") " pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.970087 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrl7s\" (UniqueName: \"kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s\") pod \"controller-manager-879f6c89f-lwgvj\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.983023 4754 request.go:700] Waited for 1.940934111s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.985137 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 11 03:08:12 crc kubenswrapper[4754]: I1011 03:08:12.986516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htngs\" (UniqueName: \"kubernetes.io/projected/6fe7f8e1-0b7e-44df-820a-924402b9b889-kube-api-access-htngs\") pod \"console-operator-58897d9998-h4qb6\" (UID: \"6fe7f8e1-0b7e-44df-820a-924402b9b889\") " pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.022173 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.024449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.025149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqkcz\" (UniqueName: \"kubernetes.io/projected/8e720967-ded2-443d-a9ae-dadbeebebdaa-kube-api-access-vqkcz\") pod \"machine-api-operator-5694c8668f-dhcrp\" (UID: \"8e720967-ded2-443d-a9ae-dadbeebebdaa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.032520 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.063666 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.063857 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-464tm\" (UniqueName: \"kubernetes.io/projected/aeac3b86-721d-47c3-99a4-523e11d3b89b-kube-api-access-464tm\") pod \"etcd-operator-b45778765-bppnf\" (UID: \"aeac3b86-721d-47c3-99a4-523e11d3b89b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.078279 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.092323 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.092898 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.097599 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.101291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.106585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq69t\" (UniqueName: \"kubernetes.io/projected/f6e20429-5a1a-469e-9826-3d5561f8c627-kube-api-access-wq69t\") pod \"authentication-operator-69f744f599-6bbxp\" (UID: \"f6e20429-5a1a-469e-9826-3d5561f8c627\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.115957 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.118247 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.131653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qf2l\" (UniqueName: \"kubernetes.io/projected/b3970eb8-365e-4236-95b0-d556e52eb675-kube-api-access-7qf2l\") pod \"openshift-controller-manager-operator-756b6f6bc6-7zsmt\" (UID: \"b3970eb8-365e-4236-95b0-d556e52eb675\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.138978 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.141702 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hnzg\" (UniqueName: \"kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg\") pod \"console-f9d7485db-m25hk\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.144602 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.148948 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.154364 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.161812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gks\" (UniqueName: \"kubernetes.io/projected/5d8e642c-404d-4b20-8443-7e00e75f3c1c-kube-api-access-c8gks\") pod \"machine-approver-56656f9798-57clb\" (UID: \"5d8e642c-404d-4b20-8443-7e00e75f3c1c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.175516 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.195679 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hh72\" (UniqueName: \"kubernetes.io/projected/7acf8241-6d60-494a-847d-ffab61ae4242-kube-api-access-8hh72\") pod \"openshift-config-operator-7777fb866f-7dc2c\" (UID: \"7acf8241-6d60-494a-847d-ffab61ae4242\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.199122 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" Oct 11 03:08:13 crc kubenswrapper[4754]: W1011 03:08:13.226649 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32a285a1_54aa_4ba5_8a04_9bfffe6ea0e8.slice/crio-dc96b6d42628a34415be0af5ab66deee4794d892daafce7fa3ed8578e5519d28 WatchSource:0}: Error finding container dc96b6d42628a34415be0af5ab66deee4794d892daafce7fa3ed8578e5519d28: Status 404 returned error can't find the container with id dc96b6d42628a34415be0af5ab66deee4794d892daafce7fa3ed8578e5519d28 Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.249360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272087 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66pd7\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272229 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272289 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272411 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272451 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.272479 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.273177 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:13.773148603 +0000 UTC m=+141.332093388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.285585 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.372044 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373032 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.373208 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:13.873182793 +0000 UTC m=+141.432127578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373448 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/567df776-e780-4c51-9940-b036917fadba-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373474 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-key\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373521 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnknx\" (UniqueName: \"kubernetes.io/projected/b16acfb8-265b-494e-9164-7a7c3e39d3f1-kube-api-access-dnknx\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373557 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw825\" (UniqueName: \"kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373581 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-webhook-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373681 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-srv-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373734 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e76199a1-bef9-4fa9-a6e0-2f95742c438a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/552772dc-229e-447d-a387-84d36f22cfe3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373821 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92hgk\" (UniqueName: \"kubernetes.io/projected/567df776-e780-4c51-9940-b036917fadba-kube-api-access-92hgk\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-images\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b3db5bf-018c-4126-a1da-fd37c06fc91a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373905 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-apiservice-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373933 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373951 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66pd7\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.373985 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374043 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmw2m\" (UniqueName: \"kubernetes.io/projected/10abaeef-73a2-4cde-aed6-54e882e8bc2b-kube-api-access-dmw2m\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374061 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-csi-data-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3db5bf-018c-4126-a1da-fd37c06fc91a-config\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374120 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8wmk\" (UniqueName: \"kubernetes.io/projected/d24258bb-2ed3-4fe6-8837-4dbd17616126-kube-api-access-q8wmk\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374139 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374159 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2fzz\" (UniqueName: \"kubernetes.io/projected/e76199a1-bef9-4fa9-a6e0-2f95742c438a-kube-api-access-k2fzz\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a09035b8-6688-4cdc-b7e9-2e55df584c31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374231 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10abaeef-73a2-4cde-aed6-54e882e8bc2b-config\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374273 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/14351961-a90b-46ee-97ce-37fb7d9d1a87-proxy-tls\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374317 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-cabundle\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374335 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-metrics-certs\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374353 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4kwx\" (UniqueName: \"kubernetes.io/projected/a4bdbf54-3723-4249-8fdb-3233805632bc-kube-api-access-f4kwx\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374371 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-srv-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcgch\" (UniqueName: \"kubernetes.io/projected/9b3b34e4-d020-44b5-afad-a8b1d926537f-kube-api-access-lcgch\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374408 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmt6n\" (UniqueName: \"kubernetes.io/projected/861415ec-1d7b-416f-9d98-0b327b553374-kube-api-access-gmt6n\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374459 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpk5c\" (UniqueName: \"kubernetes.io/projected/c30c3edd-14a5-45c4-a3e5-4128af4d20a7-kube-api-access-hpk5c\") pod \"migrator-59844c95c7-gfrq9\" (UID: \"c30c3edd-14a5-45c4-a3e5-4128af4d20a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374474 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-default-certificate\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a09035b8-6688-4cdc-b7e9-2e55df584c31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.374519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.376274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r9jh\" (UniqueName: \"kubernetes.io/projected/d2cb1a03-134c-4ec4-92b3-72766f16e416-kube-api-access-9r9jh\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.377988 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:13.877952494 +0000 UTC m=+141.436897279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.378423 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.380767 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhldh\" (UniqueName: \"kubernetes.io/projected/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-kube-api-access-hhldh\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.380821 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/861415ec-1d7b-416f-9d98-0b327b553374-metrics-tls\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.380847 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftwbz\" (UniqueName: \"kubernetes.io/projected/14351961-a90b-46ee-97ce-37fb7d9d1a87-kube-api-access-ftwbz\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381246 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10abaeef-73a2-4cde-aed6-54e882e8bc2b-serving-cert\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-profile-collector-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381343 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-plugins-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fce824f1-c109-441d-9fd9-5649273aafae-cert\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-stats-auth\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381730 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/567df776-e780-4c51-9940-b036917fadba-proxy-tls\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381755 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsqpv\" (UniqueName: \"kubernetes.io/projected/fce824f1-c109-441d-9fd9-5649273aafae-kube-api-access-jsqpv\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.381805 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.382907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-service-ca-bundle\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.383039 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a09035b8-6688-4cdc-b7e9-2e55df584c31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.383069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7vq6\" (UniqueName: \"kubernetes.io/projected/6504a02f-1d15-4635-b150-82cc7934abb5-kube-api-access-d7vq6\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.383284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-mountpoint-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.383326 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qgkm\" (UniqueName: \"kubernetes.io/projected/26390bf9-f736-4878-b900-d191cd3799d1-kube-api-access-7qgkm\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.383438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384615 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6504a02f-1d15-4635-b150-82cc7934abb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384738 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d24258bb-2ed3-4fe6-8837-4dbd17616126-tmpfs\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384763 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpg7q\" (UniqueName: \"kubernetes.io/projected/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-kube-api-access-hpg7q\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-socket-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384830 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-node-bootstrap-token\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384886 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-certs\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384915 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/861415ec-1d7b-416f-9d98-0b327b553374-config-volume\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384931 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jstrg\" (UniqueName: \"kubernetes.io/projected/552772dc-229e-447d-a387-84d36f22cfe3-kube-api-access-jstrg\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.384956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-registration-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.385006 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/26390bf9-f736-4878-b900-d191cd3799d1-metrics-tls\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.385212 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3db5bf-018c-4126-a1da-fd37c06fc91a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.385266 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8r52\" (UniqueName: \"kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.386306 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.388020 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.388179 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.419889 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66pd7\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.429651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.443859 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488052 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.488353 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:13.988313993 +0000 UTC m=+141.547258778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488605 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/567df776-e780-4c51-9940-b036917fadba-proxy-tls\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488632 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsqpv\" (UniqueName: \"kubernetes.io/projected/fce824f1-c109-441d-9fd9-5649273aafae-kube-api-access-jsqpv\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488655 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488676 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-service-ca-bundle\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a09035b8-6688-4cdc-b7e9-2e55df584c31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488747 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7vq6\" (UniqueName: \"kubernetes.io/projected/6504a02f-1d15-4635-b150-82cc7934abb5-kube-api-access-d7vq6\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488768 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-mountpoint-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488790 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qgkm\" (UniqueName: \"kubernetes.io/projected/26390bf9-f736-4878-b900-d191cd3799d1-kube-api-access-7qgkm\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488811 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488839 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6504a02f-1d15-4635-b150-82cc7934abb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d24258bb-2ed3-4fe6-8837-4dbd17616126-tmpfs\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488907 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpg7q\" (UniqueName: \"kubernetes.io/projected/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-kube-api-access-hpg7q\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488931 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-socket-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488949 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-node-bootstrap-token\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.488997 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-certs\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/861415ec-1d7b-416f-9d98-0b327b553374-config-volume\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jstrg\" (UniqueName: \"kubernetes.io/projected/552772dc-229e-447d-a387-84d36f22cfe3-kube-api-access-jstrg\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-registration-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/26390bf9-f736-4878-b900-d191cd3799d1-metrics-tls\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3db5bf-018c-4126-a1da-fd37c06fc91a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489130 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8r52\" (UniqueName: \"kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/567df776-e780-4c51-9940-b036917fadba-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-key\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnknx\" (UniqueName: \"kubernetes.io/projected/b16acfb8-265b-494e-9164-7a7c3e39d3f1-kube-api-access-dnknx\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489249 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw825\" (UniqueName: \"kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489271 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-webhook-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489343 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-srv-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489369 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/552772dc-229e-447d-a387-84d36f22cfe3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489418 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e76199a1-bef9-4fa9-a6e0-2f95742c438a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92hgk\" (UniqueName: \"kubernetes.io/projected/567df776-e780-4c51-9940-b036917fadba-kube-api-access-92hgk\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489493 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b3db5bf-018c-4126-a1da-fd37c06fc91a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489514 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-apiservice-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-images\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489597 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmw2m\" (UniqueName: \"kubernetes.io/projected/10abaeef-73a2-4cde-aed6-54e882e8bc2b-kube-api-access-dmw2m\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-csi-data-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489654 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3db5bf-018c-4126-a1da-fd37c06fc91a-config\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8wmk\" (UniqueName: \"kubernetes.io/projected/d24258bb-2ed3-4fe6-8837-4dbd17616126-kube-api-access-q8wmk\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489722 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2fzz\" (UniqueName: \"kubernetes.io/projected/e76199a1-bef9-4fa9-a6e0-2f95742c438a-kube-api-access-k2fzz\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489746 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a09035b8-6688-4cdc-b7e9-2e55df584c31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489826 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489847 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10abaeef-73a2-4cde-aed6-54e882e8bc2b-config\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489889 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/14351961-a90b-46ee-97ce-37fb7d9d1a87-proxy-tls\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-cabundle\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489935 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-metrics-certs\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.489985 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-srv-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcgch\" (UniqueName: \"kubernetes.io/projected/9b3b34e4-d020-44b5-afad-a8b1d926537f-kube-api-access-lcgch\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490046 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4kwx\" (UniqueName: \"kubernetes.io/projected/a4bdbf54-3723-4249-8fdb-3233805632bc-kube-api-access-f4kwx\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmt6n\" (UniqueName: \"kubernetes.io/projected/861415ec-1d7b-416f-9d98-0b327b553374-kube-api-access-gmt6n\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpk5c\" (UniqueName: \"kubernetes.io/projected/c30c3edd-14a5-45c4-a3e5-4128af4d20a7-kube-api-access-hpk5c\") pod \"migrator-59844c95c7-gfrq9\" (UID: \"c30c3edd-14a5-45c4-a3e5-4128af4d20a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490090 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/861415ec-1d7b-416f-9d98-0b327b553374-config-volume\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490135 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-default-certificate\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490157 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490175 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a09035b8-6688-4cdc-b7e9-2e55df584c31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490204 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-mountpoint-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.490653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.493155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-service-ca-bundle\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.493680 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a09035b8-6688-4cdc-b7e9-2e55df584c31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.494525 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h69hd"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.495075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-registration-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.499275 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-socket-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.501855 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d24258bb-2ed3-4fe6-8837-4dbd17616126-tmpfs\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.503676 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-csi-data-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.504344 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3db5bf-018c-4126-a1da-fd37c06fc91a-config\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.504923 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r9jh\" (UniqueName: \"kubernetes.io/projected/d2cb1a03-134c-4ec4-92b3-72766f16e416-kube-api-access-9r9jh\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhldh\" (UniqueName: \"kubernetes.io/projected/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-kube-api-access-hhldh\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/861415ec-1d7b-416f-9d98-0b327b553374-metrics-tls\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftwbz\" (UniqueName: \"kubernetes.io/projected/14351961-a90b-46ee-97ce-37fb7d9d1a87-kube-api-access-ftwbz\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505320 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10abaeef-73a2-4cde-aed6-54e882e8bc2b-serving-cert\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505401 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-profile-collector-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505473 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-plugins-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fce824f1-c109-441d-9fd9-5649273aafae-cert\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-stats-auth\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.505000 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.507612 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.007589701 +0000 UTC m=+141.566534676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.507684 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14351961-a90b-46ee-97ce-37fb7d9d1a87-images\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.509568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/567df776-e780-4c51-9940-b036917fadba-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.510150 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a4bdbf54-3723-4249-8fdb-3233805632bc-plugins-dir\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.510226 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-node-bootstrap-token\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.511311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10abaeef-73a2-4cde-aed6-54e882e8bc2b-config\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.512056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.512369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-cabundle\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.512772 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6504a02f-1d15-4635-b150-82cc7934abb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.514310 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.519627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.521369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/567df776-e780-4c51-9940-b036917fadba-proxy-tls\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.525479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/26390bf9-f736-4878-b900-d191cd3799d1-metrics-tls\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.534250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3db5bf-018c-4126-a1da-fd37c06fc91a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.534362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsqpv\" (UniqueName: \"kubernetes.io/projected/fce824f1-c109-441d-9fd9-5649273aafae-kube-api-access-jsqpv\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.534450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-metrics-certs\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.534892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a09035b8-6688-4cdc-b7e9-2e55df584c31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.536095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-stats-auth\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.536683 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e76199a1-bef9-4fa9-a6e0-2f95742c438a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.537927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10abaeef-73a2-4cde-aed6-54e882e8bc2b-serving-cert\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.538359 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-default-certificate\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.538824 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-srv-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.539489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-webhook-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.540063 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b16acfb8-265b-494e-9164-7a7c3e39d3f1-certs\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.540433 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d24258bb-2ed3-4fe6-8837-4dbd17616126-apiservice-cert\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.540812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/552772dc-229e-447d-a387-84d36f22cfe3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.541095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fce824f1-c109-441d-9fd9-5649273aafae-cert\") pod \"ingress-canary-c5dg2\" (UID: \"fce824f1-c109-441d-9fd9-5649273aafae\") " pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.541590 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b3b34e4-d020-44b5-afad-a8b1d926537f-profile-collector-cert\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.541780 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/14351961-a90b-46ee-97ce-37fb7d9d1a87-proxy-tls\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.542611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-signing-key\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.543269 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d2cb1a03-134c-4ec4-92b3-72766f16e416-srv-cert\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.544885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/861415ec-1d7b-416f-9d98-0b327b553374-metrics-tls\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.545167 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.547077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7vq6\" (UniqueName: \"kubernetes.io/projected/6504a02f-1d15-4635-b150-82cc7934abb5-kube-api-access-d7vq6\") pod \"multus-admission-controller-857f4d67dd-xn7rq\" (UID: \"6504a02f-1d15-4635-b150-82cc7934abb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.559562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qgkm\" (UniqueName: \"kubernetes.io/projected/26390bf9-f736-4878-b900-d191cd3799d1-kube-api-access-7qgkm\") pod \"dns-operator-744455d44c-vcpwr\" (UID: \"26390bf9-f736-4878-b900-d191cd3799d1\") " pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.581593 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpg7q\" (UniqueName: \"kubernetes.io/projected/ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7-kube-api-access-hpg7q\") pod \"service-ca-9c57cc56f-clhf8\" (UID: \"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7\") " pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.606664 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.607156 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.107135644 +0000 UTC m=+141.666080429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.607653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmw2m\" (UniqueName: \"kubernetes.io/projected/10abaeef-73a2-4cde-aed6-54e882e8bc2b-kube-api-access-dmw2m\") pod \"service-ca-operator-777779d784-cdmps\" (UID: \"10abaeef-73a2-4cde-aed6-54e882e8bc2b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.633156 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jstrg\" (UniqueName: \"kubernetes.io/projected/552772dc-229e-447d-a387-84d36f22cfe3-kube-api-access-jstrg\") pod \"control-plane-machine-set-operator-78cbb6b69f-xcwj8\" (UID: \"552772dc-229e-447d-a387-84d36f22cfe3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.647598 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.647919 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-c5dg2" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.660828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a09035b8-6688-4cdc-b7e9-2e55df584c31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lt575\" (UID: \"a09035b8-6688-4cdc-b7e9-2e55df584c31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.666363 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bppnf"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.675482 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4qb6"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.677598 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8wmk\" (UniqueName: \"kubernetes.io/projected/d24258bb-2ed3-4fe6-8837-4dbd17616126-kube-api-access-q8wmk\") pod \"packageserver-d55dfcdfc-94r9q\" (UID: \"d24258bb-2ed3-4fe6-8837-4dbd17616126\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.706556 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.714936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2fzz\" (UniqueName: \"kubernetes.io/projected/e76199a1-bef9-4fa9-a6e0-2f95742c438a-kube-api-access-k2fzz\") pod \"package-server-manager-789f6589d5-nbpp4\" (UID: \"e76199a1-bef9-4fa9-a6e0-2f95742c438a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.717946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.718577 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.218558321 +0000 UTC m=+141.777503106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.725234 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b3db5bf-018c-4126-a1da-fd37c06fc91a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg5bz\" (UID: \"8b3db5bf-018c-4126-a1da-fd37c06fc91a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.729267 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.730366 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8r52\" (UniqueName: \"kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52\") pod \"marketplace-operator-79b997595-w5w2q\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.740237 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.755878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcgch\" (UniqueName: \"kubernetes.io/projected/9b3b34e4-d020-44b5-afad-a8b1d926537f-kube-api-access-lcgch\") pod \"catalog-operator-68c6474976-6hklc\" (UID: \"9b3b34e4-d020-44b5-afad-a8b1d926537f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.760751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4kwx\" (UniqueName: \"kubernetes.io/projected/a4bdbf54-3723-4249-8fdb-3233805632bc-kube-api-access-f4kwx\") pod \"csi-hostpathplugin-dtnsb\" (UID: \"a4bdbf54-3723-4249-8fdb-3233805632bc\") " pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.780737 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmt6n\" (UniqueName: \"kubernetes.io/projected/861415ec-1d7b-416f-9d98-0b327b553374-kube-api-access-gmt6n\") pod \"dns-default-kl4ld\" (UID: \"861415ec-1d7b-416f-9d98-0b327b553374\") " pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: W1011 03:08:13.788425 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01727067_3e38_424a_9adc_f7ffbe5e32e7.slice/crio-feca07a891a5ba1cf9ffac62de5f3f9c494ca4dac8dee3e5ec513d862a4aa385 WatchSource:0}: Error finding container feca07a891a5ba1cf9ffac62de5f3f9c494ca4dac8dee3e5ec513d862a4aa385: Status 404 returned error can't find the container with id feca07a891a5ba1cf9ffac62de5f3f9c494ca4dac8dee3e5ec513d862a4aa385 Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.799740 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpk5c\" (UniqueName: \"kubernetes.io/projected/c30c3edd-14a5-45c4-a3e5-4128af4d20a7-kube-api-access-hpk5c\") pod \"migrator-59844c95c7-gfrq9\" (UID: \"c30c3edd-14a5-45c4-a3e5-4128af4d20a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.813642 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.819569 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.819779 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.319738323 +0000 UTC m=+141.878683098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.823060 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.829018 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.831828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftwbz\" (UniqueName: \"kubernetes.io/projected/14351961-a90b-46ee-97ce-37fb7d9d1a87-kube-api-access-ftwbz\") pod \"machine-config-operator-74547568cd-d2s8l\" (UID: \"14351961-a90b-46ee-97ce-37fb7d9d1a87\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.839349 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.846340 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.857300 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.859849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r9jh\" (UniqueName: \"kubernetes.io/projected/d2cb1a03-134c-4ec4-92b3-72766f16e416-kube-api-access-9r9jh\") pod \"olm-operator-6b444d44fb-wmpjq\" (UID: \"d2cb1a03-134c-4ec4-92b3-72766f16e416\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.868017 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.872790 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.877258 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h69hd" event={"ID":"9b8d25f2-185a-454a-9601-10f51d68412b","Type":"ContainerStarted","Data":"c1ba3dcd85299ed9b790a77bbae18f5e294066a4b054d90d15c64d068f970899"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.885647 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" event={"ID":"ce0cc42e-e539-4597-a986-0cdf90611c82","Type":"ContainerStarted","Data":"8906040f6cd37808eac44bb8a57144724c39c2e2b9c66517b91fcffcdeada1c9"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.886663 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" event={"ID":"ce0cc42e-e539-4597-a986-0cdf90611c82","Type":"ContainerStarted","Data":"2a6c013cfd1360b6ca39975910508401810b7425b8e88d38992d5758be754fe4"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.887171 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhldh\" (UniqueName: \"kubernetes.io/projected/7fc4ea3d-de59-4461-b58d-e74cac57dcf0-kube-api-access-hhldh\") pod \"router-default-5444994796-wq6lw\" (UID: \"7fc4ea3d-de59-4461-b58d-e74cac57dcf0\") " pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.890929 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" event={"ID":"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8","Type":"ContainerStarted","Data":"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.891003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" event={"ID":"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8","Type":"ContainerStarted","Data":"dc96b6d42628a34415be0af5ab66deee4794d892daafce7fa3ed8578e5519d28"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.891683 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.895486 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" event={"ID":"424e4796-34b6-4b66-94eb-a3c2918667de","Type":"ContainerStarted","Data":"65afcdbdb6b157b9a753a415b7680db82850a6ddd5be368a2f18933018a295f3"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.897561 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.900724 4754 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-9cc52 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.900775 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.906098 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" event={"ID":"5d8e642c-404d-4b20-8443-7e00e75f3c1c","Type":"ContainerStarted","Data":"0dcd2879470521c7597d8e9f6f3b28d4c359e76fc8bea4a5c45c0acb53443f69"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.906428 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.907892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw825\" (UniqueName: \"kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825\") pod \"collect-profiles-29335860-pv7qw\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.910772 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.911735 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" event={"ID":"6fe7f8e1-0b7e-44df-820a-924402b9b889","Type":"ContainerStarted","Data":"727319546de5761078a91a91b0a4306dd14705975d0a266b0fddedfb83416e18"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.914709 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" event={"ID":"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5","Type":"ContainerStarted","Data":"3b8b5872b763958c803417a1281845e793b0e85537ff9dd3869e804906f34d2d"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.914756 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" event={"ID":"55eb6e2b-72d1-40bd-b0d9-7d1a2218bba5","Type":"ContainerStarted","Data":"4ae0e57bf052c9119aced8b7160f2954dfe97a1c6b9b0c272f858af474b77baf"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.921146 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:13 crc kubenswrapper[4754]: E1011 03:08:13.921602 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.421585199 +0000 UTC m=+141.980529984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.921885 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" event={"ID":"aeac3b86-721d-47c3-99a4-523e11d3b89b","Type":"ContainerStarted","Data":"b9d98807dfe54bc139cddf0ec04b68a48fafca43c83aa41d0bed4632aeb572de"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.925710 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnknx\" (UniqueName: \"kubernetes.io/projected/b16acfb8-265b-494e-9164-7a7c3e39d3f1-kube-api-access-dnknx\") pod \"machine-config-server-cxhgj\" (UID: \"b16acfb8-265b-494e-9164-7a7c3e39d3f1\") " pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.925948 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" event={"ID":"01727067-3e38-424a-9adc-f7ffbe5e32e7","Type":"ContainerStarted","Data":"feca07a891a5ba1cf9ffac62de5f3f9c494ca4dac8dee3e5ec513d862a4aa385"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.929066 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.931895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" event={"ID":"53ad7084-a584-4a86-8abd-d6419f9072c1","Type":"ContainerStarted","Data":"ca38b9c4b767c3c9bd5777828af7a8962aeb00ea27f8c704a9429ef9b990f236"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.931930 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" event={"ID":"53ad7084-a584-4a86-8abd-d6419f9072c1","Type":"ContainerStarted","Data":"e9deedd541df98d0026c161666c6592cc43ccf9b90f6e7a37dde8ad6bc036cf4"} Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.942291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cxhgj" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.943358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7"] Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.943381 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92hgk\" (UniqueName: \"kubernetes.io/projected/567df776-e780-4c51-9940-b036917fadba-kube-api-access-92hgk\") pod \"machine-config-controller-84d6567774-bl8xw\" (UID: \"567df776-e780-4c51-9940-b036917fadba\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.948398 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.957122 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" Oct 11 03:08:13 crc kubenswrapper[4754]: I1011 03:08:13.960863 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.022260 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.024063 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.524023995 +0000 UTC m=+142.082968780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.051123 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-c5dg2"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.055853 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.057198 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-42nc2"] Oct 11 03:08:14 crc kubenswrapper[4754]: W1011 03:08:14.065436 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda591a71f_4786_44ca_ba67_f289f2118e79.slice/crio-e793775f9c473f2435d6df4ad9b14d3e60c500b0557570f4668301f7802538b3 WatchSource:0}: Error finding container e793775f9c473f2435d6df4ad9b14d3e60c500b0557570f4668301f7802538b3: Status 404 returned error can't find the container with id e793775f9c473f2435d6df4ad9b14d3e60c500b0557570f4668301f7802538b3 Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.076040 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dhcrp"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.077649 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.086426 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.099323 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vcpwr"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.115419 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.124063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.124231 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6bbxp"] Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.124527 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.624509942 +0000 UTC m=+142.183454737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.125917 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.131041 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c"] Oct 11 03:08:14 crc kubenswrapper[4754]: W1011 03:08:14.139272 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e720967_ded2_443d_a9ae_dadbeebebdaa.slice/crio-3606d9f35313548c46e0559cdba1d261e2fb59486b9c6087814ad0bfb2b2ccef WatchSource:0}: Error finding container 3606d9f35313548c46e0559cdba1d261e2fb59486b9c6087814ad0bfb2b2ccef: Status 404 returned error can't find the container with id 3606d9f35313548c46e0559cdba1d261e2fb59486b9c6087814ad0bfb2b2ccef Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.151622 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.188825 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.224739 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.225113 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.725092312 +0000 UTC m=+142.284037097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.300421 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.328049 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.328082 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz"] Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.329378 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.829352284 +0000 UTC m=+142.388297069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.366133 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575"] Oct 11 03:08:14 crc kubenswrapper[4754]: W1011 03:08:14.419051 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4e1c4b8_e9f1_479d_805f_4760ce24ceb7.slice/crio-d0c1a2fe3243ea290af96dea6b91229b00da1474f0c25b5c4ec216e3d3b48b41 WatchSource:0}: Error finding container d0c1a2fe3243ea290af96dea6b91229b00da1474f0c25b5c4ec216e3d3b48b41: Status 404 returned error can't find the container with id d0c1a2fe3243ea290af96dea6b91229b00da1474f0c25b5c4ec216e3d3b48b41 Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.428801 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.429204 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.929053873 +0000 UTC m=+142.487998658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.429330 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.430506 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:14.930489344 +0000 UTC m=+142.489434139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.534791 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.535722 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.035691839 +0000 UTC m=+142.594636654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.636577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.638062 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.138046853 +0000 UTC m=+142.696991638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.684854 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xn7rq"] Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.741095 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.741324 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.241304178 +0000 UTC m=+142.800248963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.741566 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.744897 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.241973442 +0000 UTC m=+142.800918227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.809653 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgb7f" podStartSLOduration=117.809629537 podStartE2EDuration="1m57.809629537s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:14.758587255 +0000 UTC m=+142.317532050" watchObservedRunningTime="2025-10-11 03:08:14.809629537 +0000 UTC m=+142.368574332" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.818041 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8"] Oct 11 03:08:14 crc kubenswrapper[4754]: W1011 03:08:14.820631 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb16acfb8_265b_494e_9164_7a7c3e39d3f1.slice/crio-1925345e5766cb457c70c58eeaef6754c9e7af6cbeb94c9632a99f9687a10b46 WatchSource:0}: Error finding container 1925345e5766cb457c70c58eeaef6754c9e7af6cbeb94c9632a99f9687a10b46: Status 404 returned error can't find the container with id 1925345e5766cb457c70c58eeaef6754c9e7af6cbeb94c9632a99f9687a10b46 Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.842517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.842877 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.342862674 +0000 UTC m=+142.901807459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.943813 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:14 crc kubenswrapper[4754]: E1011 03:08:14.944307 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.444295144 +0000 UTC m=+143.003239929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.969919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" event={"ID":"424e4796-34b6-4b66-94eb-a3c2918667de","Type":"ContainerStarted","Data":"a683cd659d0028fe136742961920c6ce02877ee65425def649745c77d6818b0c"} Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.983298 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" event={"ID":"01727067-3e38-424a-9adc-f7ffbe5e32e7","Type":"ContainerStarted","Data":"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552"} Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.984627 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.986455 4754 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-v89kr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" start-of-body= Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.986506 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.987641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" event={"ID":"6504a02f-1d15-4635-b150-82cc7934abb5","Type":"ContainerStarted","Data":"403b2757c477da3508ab63965a97b5872cd66e721bb7fe405f338a2b275ef8ee"} Oct 11 03:08:14 crc kubenswrapper[4754]: I1011 03:08:14.994763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" event={"ID":"a09035b8-6688-4cdc-b7e9-2e55df584c31","Type":"ContainerStarted","Data":"918312724ee1e23b46e52bcbd60d7a3b42b1badf6dab1437e0e9079f9b1e784c"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.000892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" event={"ID":"8e720967-ded2-443d-a9ae-dadbeebebdaa","Type":"ContainerStarted","Data":"3606d9f35313548c46e0559cdba1d261e2fb59486b9c6087814ad0bfb2b2ccef"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.009721 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h69hd" event={"ID":"9b8d25f2-185a-454a-9601-10f51d68412b","Type":"ContainerStarted","Data":"3b7f90591426ac0299dceb4c4d2a89e598eb99056dada0e580e6b03c7fca07e5"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.010502 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.016562 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-h69hd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.016622 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h69hd" podUID="9b8d25f2-185a-454a-9601-10f51d68412b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.018901 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cxhgj" event={"ID":"b16acfb8-265b-494e-9164-7a7c3e39d3f1","Type":"ContainerStarted","Data":"1925345e5766cb457c70c58eeaef6754c9e7af6cbeb94c9632a99f9687a10b46"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.022688 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" event={"ID":"f6e20429-5a1a-469e-9826-3d5561f8c627","Type":"ContainerStarted","Data":"97002f244e9ba42e3447c482e9cdb8e89a32e0f61e0a62e5a4f2a53320915a89"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.033295 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" event={"ID":"8b3db5bf-018c-4126-a1da-fd37c06fc91a","Type":"ContainerStarted","Data":"f5095f40afaaa944a4d7d7b0540d3f278528f8e8219ab0c7490852c74de55fb1"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.047313 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.047693 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.547678444 +0000 UTC m=+143.106623229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.067360 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" event={"ID":"b3970eb8-365e-4236-95b0-d556e52eb675","Type":"ContainerStarted","Data":"0a1c2170b390db0ed887d8f5d1550dc0e747b3b28a4e4edf78dd9364ba517e26"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.078631 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ntpp8" podStartSLOduration=118.078612169 podStartE2EDuration="1m58.078612169s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.078402271 +0000 UTC m=+142.637347056" watchObservedRunningTime="2025-10-11 03:08:15.078612169 +0000 UTC m=+142.637556954" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.148597 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.153036 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.653021035 +0000 UTC m=+143.211965820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.174896 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" event={"ID":"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7","Type":"ContainerStarted","Data":"d0c1a2fe3243ea290af96dea6b91229b00da1474f0c25b5c4ec216e3d3b48b41"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.174950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cdmps"] Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.181818 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" event={"ID":"26390bf9-f736-4878-b900-d191cd3799d1","Type":"ContainerStarted","Data":"c309747d8cade906e856de959a3f1f223f32bab760759aed1634d8423e67e249"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.190034 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-c5dg2" event={"ID":"fce824f1-c109-441d-9fd9-5649273aafae","Type":"ContainerStarted","Data":"b0c57c5f0af2b4d27812a7381e777cac525d48fc813373c36871de4ec378ab5e"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.202585 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" podStartSLOduration=118.202560063 podStartE2EDuration="1m58.202560063s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.161741066 +0000 UTC m=+142.720685861" watchObservedRunningTime="2025-10-11 03:08:15.202560063 +0000 UTC m=+142.761504848" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.219766 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" event={"ID":"2243842d-42a7-446c-a5a9-7d57d6fce732","Type":"ContainerStarted","Data":"994bb4a55715c852b2da6807d315f7fe1e73ff950ef04f55ebc206a2c992d4b0"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.226432 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" event={"ID":"53ad7084-a584-4a86-8abd-d6419f9072c1","Type":"ContainerStarted","Data":"9aee1453f72711831c20f8e042e2f72e48a26fb06f27004d0d6d95b1120ecf06"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.230457 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" event={"ID":"f9ca86b2-12d2-4285-a892-6b85beb92b72","Type":"ContainerStarted","Data":"89d00fa68974edc5a3feacefd0c180a194985bf3b1557c945e49a04bd4cd6c41"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.234256 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wq6lw" event={"ID":"7fc4ea3d-de59-4461-b58d-e74cac57dcf0","Type":"ContainerStarted","Data":"4b6e1a124b151c547d010244d18bd6054805cda9808a14731115dc6cd1995deb"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.239750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" event={"ID":"7acf8241-6d60-494a-847d-ffab61ae4242","Type":"ContainerStarted","Data":"1a19d048d8d2d39659f548f5603b06b1f31c81b64bc74edacbfbc121300d631b"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.248379 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" event={"ID":"a591a71f-4786-44ca-ba67-f289f2118e79","Type":"ContainerStarted","Data":"e793775f9c473f2435d6df4ad9b14d3e60c500b0557570f4668301f7802538b3"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.249515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.250384 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.750360069 +0000 UTC m=+143.309304854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.253572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" event={"ID":"5d8e642c-404d-4b20-8443-7e00e75f3c1c","Type":"ContainerStarted","Data":"39c90d6c2241653bbc0d28fc965a11a55cabef02caf12af72483b6dc8c753ebc"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.264898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" event={"ID":"aeac3b86-721d-47c3-99a4-523e11d3b89b","Type":"ContainerStarted","Data":"1d20b7b1fca4fc76018611aad265c9c73d660a94c7844ace897069a9c1a15aa2"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.272846 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" event={"ID":"d24258bb-2ed3-4fe6-8837-4dbd17616126","Type":"ContainerStarted","Data":"1186b9455262762f73096346abc86e6466902a8f5a12e6485e134dcac7d15acc"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.281875 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m25hk" event={"ID":"36eab906-00ee-4a62-b0c1-85f3daccb0d8","Type":"ContainerStarted","Data":"21214b6adfcfa8e11872a43f61132659fd341b9ec5dc03d03ea66ecf94ad13cd"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.286174 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" event={"ID":"1b0bf758-fa24-4d06-a7bc-52f8e1c05919","Type":"ContainerStarted","Data":"a677f41a1294fa1da12e67f868e1255e02d44acee4994ab7243f243f699f44c3"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.309903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" event={"ID":"6fe7f8e1-0b7e-44df-820a-924402b9b889","Type":"ContainerStarted","Data":"a04db7220ca29c9486c9dec82207979f812889ca40a12eff6f25143e640fbd02"} Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.312307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.317941 4754 patch_prober.go:28] interesting pod/console-operator-58897d9998-h4qb6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.318096 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" podUID="6fe7f8e1-0b7e-44df-820a-924402b9b889" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.351587 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.352370 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.85235478 +0000 UTC m=+143.411299565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.455721 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.456452 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:15.956431985 +0000 UTC m=+143.515376770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.505687 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-h69hd" podStartSLOduration=118.505671122 podStartE2EDuration="1m58.505671122s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.504341775 +0000 UTC m=+143.063286550" watchObservedRunningTime="2025-10-11 03:08:15.505671122 +0000 UTC m=+143.064615907" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.506877 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-bppnf" podStartSLOduration=118.506870425 podStartE2EDuration="1m58.506870425s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.451395275 +0000 UTC m=+143.010340080" watchObservedRunningTime="2025-10-11 03:08:15.506870425 +0000 UTC m=+143.065815210" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.528032 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qhjfs" podStartSLOduration=118.52801231 podStartE2EDuration="1m58.52801231s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.527672958 +0000 UTC m=+143.086617753" watchObservedRunningTime="2025-10-11 03:08:15.52801231 +0000 UTC m=+143.086957095" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.541593 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.557607 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.559682 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.0596669 +0000 UTC m=+143.618611685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.566857 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zmxds" podStartSLOduration=118.566834666 podStartE2EDuration="1m58.566834666s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.560428467 +0000 UTC m=+143.119373252" watchObservedRunningTime="2025-10-11 03:08:15.566834666 +0000 UTC m=+143.125779451" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.601846 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" podStartSLOduration=118.601829265 podStartE2EDuration="1m58.601829265s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.601086338 +0000 UTC m=+143.160031123" watchObservedRunningTime="2025-10-11 03:08:15.601829265 +0000 UTC m=+143.160774050" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.658886 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.659374 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.159353908 +0000 UTC m=+143.718298693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.660738 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" podStartSLOduration=118.660712627 podStartE2EDuration="1m58.660712627s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:15.656717334 +0000 UTC m=+143.215662129" watchObservedRunningTime="2025-10-11 03:08:15.660712627 +0000 UTC m=+143.219657412" Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.760730 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.761325 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.261293687 +0000 UTC m=+143.820238612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.862426 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.863430 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.363406732 +0000 UTC m=+143.922351517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.890923 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw"] Oct 11 03:08:15 crc kubenswrapper[4754]: I1011 03:08:15.964294 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:15 crc kubenswrapper[4754]: E1011 03:08:15.964676 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.464663316 +0000 UTC m=+144.023608101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.062296 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.066561 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.066740 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.566710509 +0000 UTC m=+144.125655294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.073835 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dtnsb"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.169724 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.170915 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.670894577 +0000 UTC m=+144.229839362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.204165 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.217076 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.221861 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kl4ld"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.223106 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.246781 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.271071 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-clhf8"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.273934 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.274327 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.774312479 +0000 UTC m=+144.333257264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.274346 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.275756 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw"] Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.334925 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" event={"ID":"567df776-e780-4c51-9940-b036917fadba","Type":"ContainerStarted","Data":"464423c5fc940b2571a8238c3fe2b0f8b710929c550e426355934cdbeb96205c"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.337233 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m25hk" event={"ID":"36eab906-00ee-4a62-b0c1-85f3daccb0d8","Type":"ContainerStarted","Data":"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037"} Oct 11 03:08:16 crc kubenswrapper[4754]: W1011 03:08:16.340542 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3b34e4_d020_44b5_afad_a8b1d926537f.slice/crio-5a985afd077db380bf2a99e4452fa92642887ea3ee548727ce4f89e8fe686b01 WatchSource:0}: Error finding container 5a985afd077db380bf2a99e4452fa92642887ea3ee548727ce4f89e8fe686b01: Status 404 returned error can't find the container with id 5a985afd077db380bf2a99e4452fa92642887ea3ee548727ce4f89e8fe686b01 Oct 11 03:08:16 crc kubenswrapper[4754]: W1011 03:08:16.358580 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafd23974_0ea6_466b_a570_2d1d32c12a84.slice/crio-580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6 WatchSource:0}: Error finding container 580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6: Status 404 returned error can't find the container with id 580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6 Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.358876 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" event={"ID":"1b0bf758-fa24-4d06-a7bc-52f8e1c05919","Type":"ContainerStarted","Data":"fd47ea0268d4d8f27f0cef5390cae1e9a426022552e49b95d369a70d8413f1bb"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.362403 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" event={"ID":"a4bdbf54-3723-4249-8fdb-3233805632bc","Type":"ContainerStarted","Data":"b07b17ed16eaafe880f2a1b76f0c2765a8d48ac92ab64e3fa844aec19f476051"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.369262 4754 generic.go:334] "Generic (PLEG): container finished" podID="f9ca86b2-12d2-4285-a892-6b85beb92b72" containerID="cf9f1e9a263637a308e0aa41f2fc6d6289e76039092251c7e7a11e7da19f3082" exitCode=0 Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.369833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" event={"ID":"f9ca86b2-12d2-4285-a892-6b85beb92b72","Type":"ContainerDied","Data":"cf9f1e9a263637a308e0aa41f2fc6d6289e76039092251c7e7a11e7da19f3082"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.379166 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.379572 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.879560406 +0000 UTC m=+144.438505191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.380148 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m25hk" podStartSLOduration=119.380126686 podStartE2EDuration="1m59.380126686s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.376828818 +0000 UTC m=+143.935773613" watchObservedRunningTime="2025-10-11 03:08:16.380126686 +0000 UTC m=+143.939071471" Oct 11 03:08:16 crc kubenswrapper[4754]: W1011 03:08:16.381257 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14351961_a90b_46ee_97ce_37fb7d9d1a87.slice/crio-57bf6314d2438d6647e83742a2ea3aeb02e2e85baacd21dbd6e4640b852ddace WatchSource:0}: Error finding container 57bf6314d2438d6647e83742a2ea3aeb02e2e85baacd21dbd6e4640b852ddace: Status 404 returned error can't find the container with id 57bf6314d2438d6647e83742a2ea3aeb02e2e85baacd21dbd6e4640b852ddace Oct 11 03:08:16 crc kubenswrapper[4754]: W1011 03:08:16.384276 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded31fd41_aa15_4aaa_8470_ad62ea6e9ee7.slice/crio-83b2bf85551d1833d85879e4d21c3392d244bbaea27bf0ebf1793f2d3a91dab1 WatchSource:0}: Error finding container 83b2bf85551d1833d85879e4d21c3392d244bbaea27bf0ebf1793f2d3a91dab1: Status 404 returned error can't find the container with id 83b2bf85551d1833d85879e4d21c3392d244bbaea27bf0ebf1793f2d3a91dab1 Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.388027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-c5dg2" event={"ID":"fce824f1-c109-441d-9fd9-5649273aafae","Type":"ContainerStarted","Data":"7222b0ded017957bcc33627b1fa6615e13d7e8e16dba518fa6bcbc1290254236"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.440831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" event={"ID":"f6e20429-5a1a-469e-9826-3d5561f8c627","Type":"ContainerStarted","Data":"13af5c84b5a134d60557caee4f47c875a9d9e1e6ec7e876bae8759d54581dc41"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.463950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" event={"ID":"b3970eb8-365e-4236-95b0-d556e52eb675","Type":"ContainerStarted","Data":"a99b0f00aeb70d3cf8dcdd45b45c3e669683f1f01a153fe13e52cba98ebdfd55"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.472657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" event={"ID":"8e720967-ded2-443d-a9ae-dadbeebebdaa","Type":"ContainerStarted","Data":"fbc5ee7af7a9883092310109deff1f47c84e655fd2fce7b75c2d314e81f7bb9d"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.480077 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.480888 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.980840991 +0000 UTC m=+144.539785776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.481251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.486051 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:16.986037646 +0000 UTC m=+144.544982431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.522106 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-c5dg2" podStartSLOduration=6.522080373 podStartE2EDuration="6.522080373s" podCreationTimestamp="2025-10-11 03:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.504804506 +0000 UTC m=+144.063749291" watchObservedRunningTime="2025-10-11 03:08:16.522080373 +0000 UTC m=+144.081025158" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.544829 4754 generic.go:334] "Generic (PLEG): container finished" podID="2243842d-42a7-446c-a5a9-7d57d6fce732" containerID="c272da2679fd833d0aecf75313574e54ab4c7573f569c185dd6ac1cdbacabb91" exitCode=0 Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.544933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" event={"ID":"2243842d-42a7-446c-a5a9-7d57d6fce732","Type":"ContainerDied","Data":"c272da2679fd833d0aecf75313574e54ab4c7573f569c185dd6ac1cdbacabb91"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.560180 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6bbxp" podStartSLOduration=119.560158662 podStartE2EDuration="1m59.560158662s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.547458339 +0000 UTC m=+144.106403124" watchObservedRunningTime="2025-10-11 03:08:16.560158662 +0000 UTC m=+144.119103447" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.572324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" event={"ID":"a591a71f-4786-44ca-ba67-f289f2118e79","Type":"ContainerStarted","Data":"395cb7360d02a1421ec7a4c04a4703215c176988e7bef4a380ff43d05173fd1f"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.582149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.582777 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7zsmt" podStartSLOduration=119.582758379 podStartE2EDuration="1m59.582758379s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.580923893 +0000 UTC m=+144.139868688" watchObservedRunningTime="2025-10-11 03:08:16.582758379 +0000 UTC m=+144.141703164" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.593647 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.093617657 +0000 UTC m=+144.652562442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.598576 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" event={"ID":"10abaeef-73a2-4cde-aed6-54e882e8bc2b","Type":"ContainerStarted","Data":"d5892bcb0d18e19753f67420946ea35be07288dbec3b63e9dcb9c92756e9c81d"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.607228 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" event={"ID":"552772dc-229e-447d-a387-84d36f22cfe3","Type":"ContainerStarted","Data":"06b5d614a995a2f3a46b23efa6f544afa1b2f59437cf6bd16d3070a39e23e098"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.607278 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" event={"ID":"552772dc-229e-447d-a387-84d36f22cfe3","Type":"ContainerStarted","Data":"55fbcd9792637532ff450e5aa37741ed1d4bc0c5af4a1218080ae89e008d2be4"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.611207 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cxhgj" event={"ID":"b16acfb8-265b-494e-9164-7a7c3e39d3f1","Type":"ContainerStarted","Data":"89a33f04890ee4a8278867422ee6f1836899e0377520b440d848c98226f9fd81"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.620531 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" event={"ID":"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7","Type":"ContainerStarted","Data":"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.621526 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.624883 4754 generic.go:334] "Generic (PLEG): container finished" podID="7acf8241-6d60-494a-847d-ffab61ae4242" containerID="c1f7bf9b1f4b18539a8f6ba3f3ea2eb1e91e3ccd2dec2c7cb6327c2813b53a0b" exitCode=0 Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.624944 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" event={"ID":"7acf8241-6d60-494a-847d-ffab61ae4242","Type":"ContainerDied","Data":"c1f7bf9b1f4b18539a8f6ba3f3ea2eb1e91e3ccd2dec2c7cb6327c2813b53a0b"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.629063 4754 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lwgvj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.629131 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.634331 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mg6v2" podStartSLOduration=119.634299009 podStartE2EDuration="1m59.634299009s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.632229555 +0000 UTC m=+144.191174350" watchObservedRunningTime="2025-10-11 03:08:16.634299009 +0000 UTC m=+144.193243794" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.655045 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wq6lw" event={"ID":"7fc4ea3d-de59-4461-b58d-e74cac57dcf0","Type":"ContainerStarted","Data":"d5eba6af552c32058dbe38712248bb94d5560c062b80f9a7310a41e4f7ca9e8f"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.684804 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.686072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" event={"ID":"e76199a1-bef9-4fa9-a6e0-2f95742c438a","Type":"ContainerStarted","Data":"e09da5dce75f03c44e338498e9d2e67aa4671d1a17b21e47cad4ba30ff697250"} Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.686841 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.186825874 +0000 UTC m=+144.745770659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.707831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" event={"ID":"d24258bb-2ed3-4fe6-8837-4dbd17616126","Type":"ContainerStarted","Data":"d1623c88c55b6290bd518198afc780584a9416e37507b85c92c4561e2e1b7673"} Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.708125 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xcwj8" podStartSLOduration=119.708104173 podStartE2EDuration="1m59.708104173s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.707135198 +0000 UTC m=+144.266079983" watchObservedRunningTime="2025-10-11 03:08:16.708104173 +0000 UTC m=+144.267048958" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.709678 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-h69hd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.709754 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h69hd" podUID="9b8d25f2-185a-454a-9601-10f51d68412b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.722914 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.730906 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-cxhgj" podStartSLOduration=6.73042447 podStartE2EDuration="6.73042447s" podCreationTimestamp="2025-10-11 03:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.729731505 +0000 UTC m=+144.288676290" watchObservedRunningTime="2025-10-11 03:08:16.73042447 +0000 UTC m=+144.289369255" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.765632 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-h4qb6" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.786026 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" podStartSLOduration=119.786003763 podStartE2EDuration="1m59.786003763s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.757772876 +0000 UTC m=+144.316717671" watchObservedRunningTime="2025-10-11 03:08:16.786003763 +0000 UTC m=+144.344948548" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.786146 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.788603 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-wq6lw" podStartSLOduration=119.788595115 podStartE2EDuration="1m59.788595115s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.785570987 +0000 UTC m=+144.344515772" watchObservedRunningTime="2025-10-11 03:08:16.788595115 +0000 UTC m=+144.347539900" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.792546 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.292522275 +0000 UTC m=+144.851467060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.852247 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" podStartSLOduration=119.852224896 podStartE2EDuration="1m59.852224896s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:16.852142984 +0000 UTC m=+144.411087769" watchObservedRunningTime="2025-10-11 03:08:16.852224896 +0000 UTC m=+144.411169681" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.895802 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.896433 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.396415734 +0000 UTC m=+144.955360529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.948739 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.951233 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.951284 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 11 03:08:16 crc kubenswrapper[4754]: I1011 03:08:16.997263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:16 crc kubenswrapper[4754]: E1011 03:08:16.998181 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.498156525 +0000 UTC m=+145.057101310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.108218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.108930 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.608912459 +0000 UTC m=+145.167857244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.216730 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.217246 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.717213935 +0000 UTC m=+145.276158720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.217553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.218013 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.717997503 +0000 UTC m=+145.276942288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.322404 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.322903 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.822880206 +0000 UTC m=+145.381825001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.426047 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.426387 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:17.926373791 +0000 UTC m=+145.485318576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.527947 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.528624 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.02860812 +0000 UTC m=+145.587552905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.630360 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.130344541 +0000 UTC m=+145.689289316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.629995 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.722158 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" event={"ID":"2243842d-42a7-446c-a5a9-7d57d6fce732","Type":"ContainerStarted","Data":"425fcc3fa85966e866684f693ba48e14d3b8647bb824757fe60775a95ae18d0c"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.724637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" event={"ID":"d77034e9-1126-4903-8484-a99e47d2aa7f","Type":"ContainerStarted","Data":"3b7e29745a134eb2afb18b252e726099e6e9a8d1f4c650ea8059de67cacc1e2d"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.724663 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" event={"ID":"d77034e9-1126-4903-8484-a99e47d2aa7f","Type":"ContainerStarted","Data":"3f5e6b94f3a4708906952f5f7058d22b8df255d4a3afe7e230dab175fa64ae9b"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.725774 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.728393 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w5w2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.728430 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.731329 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.731567 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.231548034 +0000 UTC m=+145.790492819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.731884 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.732206 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.232190427 +0000 UTC m=+145.791135212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.732672 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kl4ld" event={"ID":"861415ec-1d7b-416f-9d98-0b327b553374","Type":"ContainerStarted","Data":"71c8bf9ead6a9d08487c918956fde8dcb767a624c41025c544b8ab5008a125fa"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.732708 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kl4ld" event={"ID":"861415ec-1d7b-416f-9d98-0b327b553374","Type":"ContainerStarted","Data":"e13a12362626098e9662f3e0ed09130f0126cbb590255b43ea83a82fbcae495b"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.744536 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" podStartSLOduration=120.744517617 podStartE2EDuration="2m0.744517617s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.743169079 +0000 UTC m=+145.302113864" watchObservedRunningTime="2025-10-11 03:08:17.744517617 +0000 UTC m=+145.303462402" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.749767 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" event={"ID":"10abaeef-73a2-4cde-aed6-54e882e8bc2b","Type":"ContainerStarted","Data":"499fece31934b1c71e1f0243e2e5854ba8e3176f0bd4a8d2db000c2fe35e0c34"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.753799 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" event={"ID":"9b3b34e4-d020-44b5-afad-a8b1d926537f","Type":"ContainerStarted","Data":"43f42ab9b39b2542975b2c38210368cadb2b0fdb02e60efe1a738a404f3afee8"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.753862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" event={"ID":"9b3b34e4-d020-44b5-afad-a8b1d926537f","Type":"ContainerStarted","Data":"5a985afd077db380bf2a99e4452fa92642887ea3ee548727ce4f89e8fe686b01"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.754360 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.756778 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" event={"ID":"26390bf9-f736-4878-b900-d191cd3799d1","Type":"ContainerStarted","Data":"dbf0c7079dadfe33e5807698504f7cb60c16bb9174d8999e2fd98d471ee6a602"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.762703 4754 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6hklc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.763193 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" podUID="9b3b34e4-d020-44b5-afad-a8b1d926537f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.766404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" event={"ID":"e76199a1-bef9-4fa9-a6e0-2f95742c438a","Type":"ContainerStarted","Data":"89b711d40cbf6c712690ca8eaac7148b50f01d4cff7d587a37f20b351882397c"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.766437 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" event={"ID":"e76199a1-bef9-4fa9-a6e0-2f95742c438a","Type":"ContainerStarted","Data":"0921a63658a195ae78cbee7d87483f7471fcef1026d0096345d0e00a0e5c5908"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.766852 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.782509 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" event={"ID":"6504a02f-1d15-4635-b150-82cc7934abb5","Type":"ContainerStarted","Data":"8c65d0870d2058fc8562385c72dfed87ead1bb26b9d9a8db69d7207b2d9ceb47"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.782775 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cdmps" podStartSLOduration=120.782757141 podStartE2EDuration="2m0.782757141s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.782472231 +0000 UTC m=+145.341417016" watchObservedRunningTime="2025-10-11 03:08:17.782757141 +0000 UTC m=+145.341701926" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.791630 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" event={"ID":"8b3db5bf-018c-4126-a1da-fd37c06fc91a","Type":"ContainerStarted","Data":"38938f970c770ddc83c1f97a6a761243a21f5e37c6518084fb6bf5d6172f43d5"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.792561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" event={"ID":"a09035b8-6688-4cdc-b7e9-2e55df584c31","Type":"ContainerStarted","Data":"ee66a9925cdcf2d92198a0c5feb83065bede9b2498f7e50a8771abaf0888e6db"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.793922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" event={"ID":"14351961-a90b-46ee-97ce-37fb7d9d1a87","Type":"ContainerStarted","Data":"9404a8bbddc2dcc3e6b78b5a2fef8c804ccb3e113bbc250e55a1990bd953db02"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.793946 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" event={"ID":"14351961-a90b-46ee-97ce-37fb7d9d1a87","Type":"ContainerStarted","Data":"55f94d5d3235b6a1af2b310da61e5e421bd25c4bb226993e584a17732fc53a41"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.793955 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" event={"ID":"14351961-a90b-46ee-97ce-37fb7d9d1a87","Type":"ContainerStarted","Data":"57bf6314d2438d6647e83742a2ea3aeb02e2e85baacd21dbd6e4640b852ddace"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.807159 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" event={"ID":"f9ca86b2-12d2-4285-a892-6b85beb92b72","Type":"ContainerStarted","Data":"f755aba783eef2c7215705d1e6b70a7d1f7596b2074923ce632d777dca62c7f1"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.814862 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" podStartSLOduration=120.814842087 podStartE2EDuration="2m0.814842087s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.812582826 +0000 UTC m=+145.371527611" watchObservedRunningTime="2025-10-11 03:08:17.814842087 +0000 UTC m=+145.373786872" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.820422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" event={"ID":"c30c3edd-14a5-45c4-a3e5-4128af4d20a7","Type":"ContainerStarted","Data":"ddb7765db1b8a96d17a91a31607b10c11646a80c71324aab521a650230d53410"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.820461 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" event={"ID":"c30c3edd-14a5-45c4-a3e5-4128af4d20a7","Type":"ContainerStarted","Data":"e739af3a8cca13d46a672e9ba1aed352a06c1a78cc50d3aab24ad9161ee2ade2"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.820471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" event={"ID":"c30c3edd-14a5-45c4-a3e5-4128af4d20a7","Type":"ContainerStarted","Data":"0d090267066434f8ca81030aefffcc52a9d069aabea8e38e09ff0a533fc0d143"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.826478 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" event={"ID":"7acf8241-6d60-494a-847d-ffab61ae4242","Type":"ContainerStarted","Data":"f0fad0b25051b2cc153989dcfeb00ad632782da2b465086f7cd078e155321bd0"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.826523 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.832762 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.833588 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.333563685 +0000 UTC m=+145.892508500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.845936 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" podStartSLOduration=120.845914466 podStartE2EDuration="2m0.845914466s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.844651771 +0000 UTC m=+145.403596556" watchObservedRunningTime="2025-10-11 03:08:17.845914466 +0000 UTC m=+145.404859251" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.858102 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" event={"ID":"8e720967-ded2-443d-a9ae-dadbeebebdaa","Type":"ContainerStarted","Data":"30c2d519746e32b7b28a9608cc32de2dbfb8ef404b53a668692251dd8e90fadc"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.864412 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" event={"ID":"1b0bf758-fa24-4d06-a7bc-52f8e1c05919","Type":"ContainerStarted","Data":"7508e76a378607e153df36262bbdb9c626a952dd3a598d4fa956caf2141e04b7"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.871243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" event={"ID":"afd23974-0ea6-466b-a570-2d1d32c12a84","Type":"ContainerStarted","Data":"703a85e9411fe67d8e4afb6675677e8ea68737ffdb850166aaf66a1b37bb99d1"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.871309 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" event={"ID":"afd23974-0ea6-466b-a570-2d1d32c12a84","Type":"ContainerStarted","Data":"580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.886348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" event={"ID":"d2cb1a03-134c-4ec4-92b3-72766f16e416","Type":"ContainerStarted","Data":"4a0644436b8e6c5a1e6a06fe9a004d3b6d62389bbbcb6744bc312afa4cba3a9c"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.886408 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" event={"ID":"d2cb1a03-134c-4ec4-92b3-72766f16e416","Type":"ContainerStarted","Data":"34125f58e1465dfd703867ece3deb74d0c0364b36fb468d8860c150fd310d65c"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.887022 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.889871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" event={"ID":"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7","Type":"ContainerStarted","Data":"15827a90a2220adfbc2aeab068e8966c4c68c60d63501d399635ee6314fa5a1c"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.889910 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" event={"ID":"ed31fd41-aa15-4aaa-8470-ad62ea6e9ee7","Type":"ContainerStarted","Data":"83b2bf85551d1833d85879e4d21c3392d244bbaea27bf0ebf1793f2d3a91dab1"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.896783 4754 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wmpjq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.896826 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" podUID="d2cb1a03-134c-4ec4-92b3-72766f16e416" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.909227 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg5bz" podStartSLOduration=120.909205195 podStartE2EDuration="2m0.909205195s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.875569664 +0000 UTC m=+145.434514449" watchObservedRunningTime="2025-10-11 03:08:17.909205195 +0000 UTC m=+145.468149980" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.910331 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" event={"ID":"5d8e642c-404d-4b20-8443-7e00e75f3c1c","Type":"ContainerStarted","Data":"355abd50faf1b9b1bbd76244c8f188daa84c0e6b05e10c8c7fa36f77cc5ff441"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.923657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" event={"ID":"567df776-e780-4c51-9940-b036917fadba","Type":"ContainerStarted","Data":"857a7d3612d4455c893941428e8c0b196522968404a00a7e630da28d3225bc10"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.923717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" event={"ID":"567df776-e780-4c51-9940-b036917fadba","Type":"ContainerStarted","Data":"dcfad1c71bea1b0c4887d79933a54f131c68373bbfb8904ef3b14314b98fbeca"} Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.925793 4754 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lwgvj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.925851 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.927571 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.935514 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.961182 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfrq9" podStartSLOduration=120.940984019 podStartE2EDuration="2m0.940984019s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.939998294 +0000 UTC m=+145.498943079" watchObservedRunningTime="2025-10-11 03:08:17.940984019 +0000 UTC m=+145.499928804" Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.962442 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" podStartSLOduration=120.962425295 podStartE2EDuration="2m0.962425295s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:17.909735144 +0000 UTC m=+145.468679929" watchObservedRunningTime="2025-10-11 03:08:17.962425295 +0000 UTC m=+145.521370080" Oct 11 03:08:17 crc kubenswrapper[4754]: E1011 03:08:17.964056 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.464023292 +0000 UTC m=+146.022968077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.970699 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:17 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:17 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:17 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:17 crc kubenswrapper[4754]: I1011 03:08:17.971152 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.026971 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lt575" podStartSLOduration=121.026925167 podStartE2EDuration="2m1.026925167s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.013860801 +0000 UTC m=+145.572805586" watchObservedRunningTime="2025-10-11 03:08:18.026925167 +0000 UTC m=+145.585869952" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.036942 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.042747 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.542721461 +0000 UTC m=+146.101666236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.056444 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" podStartSLOduration=121.05641773 podStartE2EDuration="2m1.05641773s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.050240849 +0000 UTC m=+145.609185654" watchObservedRunningTime="2025-10-11 03:08:18.05641773 +0000 UTC m=+145.615362515" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.075007 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d2s8l" podStartSLOduration=121.074980262 podStartE2EDuration="2m1.074980262s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.074492695 +0000 UTC m=+145.633437490" watchObservedRunningTime="2025-10-11 03:08:18.074980262 +0000 UTC m=+145.633925047" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.099166 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.099250 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.111633 4754 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-q5dr7 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.111715 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" podUID="f9ca86b2-12d2-4285-a892-6b85beb92b72" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.126140 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" podStartSLOduration=121.126118008 podStartE2EDuration="2m1.126118008s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.125308479 +0000 UTC m=+145.684253264" watchObservedRunningTime="2025-10-11 03:08:18.126118008 +0000 UTC m=+145.685062793" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.142034 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.142455 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.64243987 +0000 UTC m=+146.201384655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.155365 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-57clb" podStartSLOduration=121.155344841 podStartE2EDuration="2m1.155344841s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.153566267 +0000 UTC m=+145.712511052" watchObservedRunningTime="2025-10-11 03:08:18.155344841 +0000 UTC m=+145.714289626" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.177047 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" podStartSLOduration=121.177019965 podStartE2EDuration="2m1.177019965s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.176660902 +0000 UTC m=+145.735605697" watchObservedRunningTime="2025-10-11 03:08:18.177019965 +0000 UTC m=+145.735964740" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.195950 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bl8xw" podStartSLOduration=121.19592572 podStartE2EDuration="2m1.19592572s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.194277541 +0000 UTC m=+145.753222326" watchObservedRunningTime="2025-10-11 03:08:18.19592572 +0000 UTC m=+145.754870495" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.240455 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fkmrd" podStartSLOduration=121.240419398 podStartE2EDuration="2m1.240419398s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.230609718 +0000 UTC m=+145.789554503" watchObservedRunningTime="2025-10-11 03:08:18.240419398 +0000 UTC m=+145.799364183" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.243724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.244037 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.744021196 +0000 UTC m=+146.302965971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.267493 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" podStartSLOduration=121.267468063 podStartE2EDuration="2m1.267468063s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.267237315 +0000 UTC m=+145.826182100" watchObservedRunningTime="2025-10-11 03:08:18.267468063 +0000 UTC m=+145.826412848" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.302729 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-clhf8" podStartSLOduration=121.302693321 podStartE2EDuration="2m1.302693321s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.298854504 +0000 UTC m=+145.857799279" watchObservedRunningTime="2025-10-11 03:08:18.302693321 +0000 UTC m=+145.861638096" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.334461 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-dhcrp" podStartSLOduration=121.334438574 podStartE2EDuration="2m1.334438574s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.333157208 +0000 UTC m=+145.892101993" watchObservedRunningTime="2025-10-11 03:08:18.334438574 +0000 UTC m=+145.893383359" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.346350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.346829 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.846811875 +0000 UTC m=+146.405756660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.447943 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.448395 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:18.948370671 +0000 UTC m=+146.507315456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.549702 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.550181 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.050163994 +0000 UTC m=+146.609108779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.565076 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-94r9q" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.651383 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.651906 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.151876335 +0000 UTC m=+146.710821120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.753054 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.753433 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.253419309 +0000 UTC m=+146.812364094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.854727 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.854956 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.354918292 +0000 UTC m=+146.913863077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.855470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.855754 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.355747102 +0000 UTC m=+146.914691887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.931787 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" event={"ID":"26390bf9-f736-4878-b900-d191cd3799d1","Type":"ContainerStarted","Data":"23fc69ea24920b14c2f6b597b0906f3ef0543cbfd0becbdfbf1b72ebc3ae7301"} Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.940623 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kl4ld" event={"ID":"861415ec-1d7b-416f-9d98-0b327b553374","Type":"ContainerStarted","Data":"63a75c1bb4ebdfc1d17aec9b693c3f413cd0a4dcbfa978ff5f1e69682db518e8"} Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.940766 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.944193 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xn7rq" event={"ID":"6504a02f-1d15-4635-b150-82cc7934abb5","Type":"ContainerStarted","Data":"e07c0769877fdfe833e8fd66c121a05b6944fa773a3e2092c3ef235283d40d3b"} Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.947175 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" event={"ID":"a4bdbf54-3723-4249-8fdb-3233805632bc","Type":"ContainerStarted","Data":"3c6deec336016439cbde113ac37f25f2b3bae7ed7b443a19b61ba75bc2a56170"} Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.952843 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" event={"ID":"2243842d-42a7-446c-a5a9-7d57d6fce732","Type":"ContainerStarted","Data":"b4b7d8db03436c64407834664635063efad50a2cb48f9fb633221083574ff8e9"} Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.955992 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w5w2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.956062 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.976193 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:18 crc kubenswrapper[4754]: E1011 03:08:18.977146 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.477118364 +0000 UTC m=+147.036063149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.977216 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:18 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:18 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:18 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.977335 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.979920 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vcpwr" podStartSLOduration=121.979896723 podStartE2EDuration="2m1.979896723s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:18.977720706 +0000 UTC m=+146.536665491" watchObservedRunningTime="2025-10-11 03:08:18.979896723 +0000 UTC m=+146.538841508" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.987307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6hklc" Oct 11 03:08:18 crc kubenswrapper[4754]: I1011 03:08:18.999430 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.009822 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wmpjq" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.079499 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.086729 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.586710776 +0000 UTC m=+147.145655561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.124474 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" podStartSLOduration=122.124450483 podStartE2EDuration="2m2.124450483s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:19.05069228 +0000 UTC m=+146.609637065" watchObservedRunningTime="2025-10-11 03:08:19.124450483 +0000 UTC m=+146.683395268" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.125586 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kl4ld" podStartSLOduration=9.125581904 podStartE2EDuration="9.125581904s" podCreationTimestamp="2025-10-11 03:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:19.103498085 +0000 UTC m=+146.662442870" watchObservedRunningTime="2025-10-11 03:08:19.125581904 +0000 UTC m=+146.684526689" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.181492 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.181810 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.68179224 +0000 UTC m=+147.240737025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.283877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.284212 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.784198005 +0000 UTC m=+147.343142790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.389703 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.390215 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.890194669 +0000 UTC m=+147.449139454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.494296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.494980 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:19.994950268 +0000 UTC m=+147.553895053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.595515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.596273 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.096250314 +0000 UTC m=+147.655195099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.698135 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.698540 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.198523685 +0000 UTC m=+147.757468470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.799806 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.800140 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.30009344 +0000 UTC m=+147.859038235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.800383 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.800456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.800584 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.801767 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.301747469 +0000 UTC m=+147.860692254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.801977 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.811831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.902228 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.902414 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.402381041 +0000 UTC m=+147.961325816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.902916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.902942 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.903049 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:19 crc kubenswrapper[4754]: E1011 03:08:19.903679 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.403655627 +0000 UTC m=+147.962600402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.907414 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.907872 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.962150 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:19 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:19 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:19 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.962211 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.986103 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" event={"ID":"a4bdbf54-3723-4249-8fdb-3233805632bc","Type":"ContainerStarted","Data":"46435c59f2a8ade35b6775aeeaf310a667e2daf0572e0f3c3a95a1c2c03bc2e0"} Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.986160 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" event={"ID":"a4bdbf54-3723-4249-8fdb-3233805632bc","Type":"ContainerStarted","Data":"32b48aedd3d30a3bffb1ebff0dfce3bfb368204de02013c71d8b773cfad4b576"} Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.987171 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w5w2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 11 03:08:19 crc kubenswrapper[4754]: I1011 03:08:19.987245 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.007712 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.008210 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.508182968 +0000 UTC m=+148.067127763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.017257 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.029998 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.041475 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.111080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.112435 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.612414319 +0000 UTC m=+148.171359104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.212803 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.213139 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.713095792 +0000 UTC m=+148.272040577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.213210 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.213656 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.713648382 +0000 UTC m=+148.272593167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.314496 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.314743 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.814709919 +0000 UTC m=+148.373654704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.314947 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.315364 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.815353802 +0000 UTC m=+148.374298587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.422412 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.423229 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:20.923212521 +0000 UTC m=+148.482157306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.527753 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.528281 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.028266931 +0000 UTC m=+148.587211716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.628445 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.628746 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.128730728 +0000 UTC m=+148.687675513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.677891 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.678840 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.683660 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.730685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.731059 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.23104703 +0000 UTC m=+148.789991815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.803154 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.832824 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.833075 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q56ll\" (UniqueName: \"kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.833111 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.833141 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.833267 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.333250948 +0000 UTC m=+148.892195733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.907587 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.908693 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.920307 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.922371 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.935025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.935071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.935109 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.935213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q56ll\" (UniqueName: \"kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.936031 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.936331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:20 crc kubenswrapper[4754]: E1011 03:08:20.936624 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.436597997 +0000 UTC m=+148.995542952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.967795 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:20 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:20 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:20 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:20 crc kubenswrapper[4754]: I1011 03:08:20.967880 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.008300 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q56ll\" (UniqueName: \"kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll\") pod \"community-operators-z6p8f\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.025736 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"08fd8475cefc13f6664078a85897d43c2d995897684921352407cc85fe33bf79"} Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.037567 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:21 crc kubenswrapper[4754]: E1011 03:08:21.037683 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.537663675 +0000 UTC m=+149.096608460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.037850 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0946de8183e0496a908946022964e693721ed377db3eae6d8ed687eb97fa4bbe"} Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.037579 4754 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.037981 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.038032 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87wgk\" (UniqueName: \"kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.038085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.038113 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: E1011 03:08:21.038505 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.538492184 +0000 UTC m=+149.097436969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.086406 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" event={"ID":"a4bdbf54-3723-4249-8fdb-3233805632bc","Type":"ContainerStarted","Data":"6ca1e848b779c069c10ae661948fc0e5a94a34bcf8939d21bef08b57c7865e58"} Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.086545 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.124618 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.126111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.138989 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.139251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.139325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.139361 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87wgk\" (UniqueName: \"kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: E1011 03:08:21.139677 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.639659595 +0000 UTC m=+149.198604380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.140052 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.140277 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.147837 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.198202 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87wgk\" (UniqueName: \"kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk\") pod \"certified-operators-bh57p\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.208735 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-dtnsb" podStartSLOduration=11.20871451 podStartE2EDuration="11.20871451s" podCreationTimestamp="2025-10-11 03:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:21.160000621 +0000 UTC m=+148.718945406" watchObservedRunningTime="2025-10-11 03:08:21.20871451 +0000 UTC m=+148.767659295" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.231191 4754 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-11T03:08:21.037976396Z","Handler":null,"Name":""} Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.245730 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crmg5\" (UniqueName: \"kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.245784 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.245844 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.245887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: E1011 03:08:21.247262 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:08:21.747245886 +0000 UTC m=+149.306190671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx5qd" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.268453 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.319612 4754 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.319679 4754 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.338404 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.350033 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.350337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.350412 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crmg5\" (UniqueName: \"kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.350476 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.350929 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.351334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.351701 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.366378 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.379935 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.416516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crmg5\" (UniqueName: \"kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5\") pod \"community-operators-gxp22\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.455040 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.455136 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trngh\" (UniqueName: \"kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.455192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.455267 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.463571 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.524529 4754 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.524580 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.557892 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.557941 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.557999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trngh\" (UniqueName: \"kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.558750 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.558769 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.593421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trngh\" (UniqueName: \"kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh\") pod \"certified-operators-7nb8q\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.666042 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx5qd\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.673125 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.816154 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.890926 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.931668 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.955213 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:21 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:21 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:21 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:21 crc kubenswrapper[4754]: I1011 03:08:21.955278 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.054849 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:08:22 crc kubenswrapper[4754]: W1011 03:08:22.063253 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf76dd9_1d7b_437b_9095_f5f043fce7f3.slice/crio-065ee6454e2701abdb433efbd333fa84c8620a4e63fbb36c93c4aed18a7daeda WatchSource:0}: Error finding container 065ee6454e2701abdb433efbd333fa84c8620a4e63fbb36c93c4aed18a7daeda: Status 404 returned error can't find the container with id 065ee6454e2701abdb433efbd333fa84c8620a4e63fbb36c93c4aed18a7daeda Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.161366 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerStarted","Data":"5c9cc2c32991ff5941c4787064be4df6007c92d407578159a0b891fb3e02fb7b"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.167434 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.175603 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8ee0f24b995a973a7d991473d7b6a5c6843ba0ea49d017372746eb58a6f944b0"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.175654 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4e0c74c4ba2fff45b957a6aa33e99707cae511885513ed1d9661f6105da180b7"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.175877 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.198551 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerStarted","Data":"065ee6454e2701abdb433efbd333fa84c8620a4e63fbb36c93c4aed18a7daeda"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.226231 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"93991634a5d9866f83193f41f7583c35d9edb1361fd4e363c77010950980286a"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.266265 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"88de8d2a56c36cc4bbf837b202bdde241c4fbd98a5abd5417e87962c4df60a08"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.286485 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerStarted","Data":"c0a85b75423dffe538cbddc702bc2b07443e5d60e4f4e8c55620045c77d82bca"} Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.387337 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7dc2c" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.464726 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.874888 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.876380 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.878403 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.905365 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.911188 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.912026 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.914578 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.915713 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.930629 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.953478 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:22 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:22 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:22 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.954021 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.958051 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-h69hd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.958085 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-h69hd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.958132 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h69hd" podUID="9b8d25f2-185a-454a-9601-10f51d68412b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 11 03:08:22 crc kubenswrapper[4754]: I1011 03:08:22.958144 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h69hd" podUID="9b8d25f2-185a-454a-9601-10f51d68412b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.000729 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwqtd\" (UniqueName: \"kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.000807 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.000912 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.001021 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.001188 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.092953 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.102948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwqtd\" (UniqueName: \"kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103172 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.103742 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.104628 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.110082 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q5dr7" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.124864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.125206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwqtd\" (UniqueName: \"kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd\") pod \"redhat-marketplace-r727p\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.145643 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.146103 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.147680 4754 patch_prober.go:28] interesting pod/console-f9d7485db-m25hk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.147752 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m25hk" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.149794 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.149850 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.163921 4754 patch_prober.go:28] interesting pod/apiserver-76f77b778f-42nc2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]log ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]etcd ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/generic-apiserver-start-informers ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/max-in-flight-filter ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 11 03:08:23 crc kubenswrapper[4754]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 11 03:08:23 crc kubenswrapper[4754]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/project.openshift.io-projectcache ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/openshift.io-startinformers ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 11 03:08:23 crc kubenswrapper[4754]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 11 03:08:23 crc kubenswrapper[4754]: livez check failed Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.164493 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" podUID="2243842d-42a7-446c-a5a9-7d57d6fce732" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.190130 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.225075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.283558 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.287043 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.304520 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.325559 4754 generic.go:334] "Generic (PLEG): container finished" podID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerID="73a69642a26637111b462b6c4bcb90a12be59d70d62e31ec1d579c4e6f66ee3e" exitCode=0 Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.325702 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerDied","Data":"73a69642a26637111b462b6c4bcb90a12be59d70d62e31ec1d579c4e6f66ee3e"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.327334 4754 generic.go:334] "Generic (PLEG): container finished" podID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerID="6ba6b12a8f6d2d3e2af75dee471a100e55c9d587b02ba765030d49685a95d3a7" exitCode=0 Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.327390 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerDied","Data":"6ba6b12a8f6d2d3e2af75dee471a100e55c9d587b02ba765030d49685a95d3a7"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.338633 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.345045 4754 generic.go:334] "Generic (PLEG): container finished" podID="afd23974-0ea6-466b-a570-2d1d32c12a84" containerID="703a85e9411fe67d8e4afb6675677e8ea68737ffdb850166aaf66a1b37bb99d1" exitCode=0 Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.345153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" event={"ID":"afd23974-0ea6-466b-a570-2d1d32c12a84","Type":"ContainerDied","Data":"703a85e9411fe67d8e4afb6675677e8ea68737ffdb850166aaf66a1b37bb99d1"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.351044 4754 generic.go:334] "Generic (PLEG): container finished" podID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerID="be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9" exitCode=0 Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.351116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerDied","Data":"be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.403773 4754 generic.go:334] "Generic (PLEG): container finished" podID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerID="7b8b911db2ece96c2f17bd8a7ba81fed20a5e044fe4927c4d483b52a0a636900" exitCode=0 Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.404556 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerDied","Data":"7b8b911db2ece96c2f17bd8a7ba81fed20a5e044fe4927c4d483b52a0a636900"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.404652 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerStarted","Data":"0a99827a4fb635d1aa77474037a61f2822e60cde0925315534dfdd4e3184c552"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.407005 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.407058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px6td\" (UniqueName: \"kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.407116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.451902 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" event={"ID":"4eefc33f-2e4f-4bd2-9866-2fc103a44135","Type":"ContainerStarted","Data":"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.451984 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.452004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" event={"ID":"4eefc33f-2e4f-4bd2-9866-2fc103a44135","Type":"ContainerStarted","Data":"38e360f4862989cf32df1b071d4d48b257987bb45071ef0e9387c6ab4fe0455a"} Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.520502 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px6td\" (UniqueName: \"kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.520619 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.520752 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.522834 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.537131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.583103 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.586613 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.587057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.599364 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px6td\" (UniqueName: \"kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td\") pod \"redhat-marketplace-4sn62\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.600380 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.608189 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.619800 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.655275 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" podStartSLOduration=126.65524706 podStartE2EDuration="2m6.65524706s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:23.611067563 +0000 UTC m=+151.170012358" watchObservedRunningTime="2025-10-11 03:08:23.65524706 +0000 UTC m=+151.214191845" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.711482 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.722849 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.723376 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.824733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.824796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.826799 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.846505 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.872520 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.874616 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.888086 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.891712 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.913203 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.915555 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.926171 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.933636 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gxp7\" (UniqueName: \"kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.933722 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.933927 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.954486 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.961582 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:23 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:23 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:23 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.961632 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:23 crc kubenswrapper[4754]: I1011 03:08:23.967929 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.035990 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.036560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gxp7\" (UniqueName: \"kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.036576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.036618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.038182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.093035 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gxp7\" (UniqueName: \"kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7\") pod \"redhat-operators-hvc2g\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.218255 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:08:24 crc kubenswrapper[4754]: W1011 03:08:24.264804 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poded6df735_dd22_44b6_bcb0_c91199885a2f.slice/crio-2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9 WatchSource:0}: Error finding container 2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9: Status 404 returned error can't find the container with id 2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9 Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.268970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.269945 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.271495 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.282322 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.350354 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.350439 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.350550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthrc\" (UniqueName: \"kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.452285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.455259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.455557 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthrc\" (UniqueName: \"kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.455635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.455993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.481739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthrc\" (UniqueName: \"kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc\") pod \"redhat-operators-8xmn8\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.484719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ddc52c0-5340-40e5-8a26-3b6188a72b70","Type":"ContainerStarted","Data":"b3a8f225c2464c449bed982166c16bef02416b17909fdc24bce1b5e8faa9d424"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.484774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ddc52c0-5340-40e5-8a26-3b6188a72b70","Type":"ContainerStarted","Data":"64e378df92b1f2ed33a21bee36438b1d77f6154da8fa342319b47c7da5790aae"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.522044 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.527455 4754 generic.go:334] "Generic (PLEG): container finished" podID="49c5c869-51bd-442b-8567-98cfd45c4765" containerID="996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281" exitCode=0 Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.527567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerDied","Data":"996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.527597 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerStarted","Data":"10ca9ff453c639290bbaf753c666fdc9321ee8efb76deb9c250a3c42d39b10eb"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.532771 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerID="027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144" exitCode=0 Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.532918 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerDied","Data":"027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.532982 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerStarted","Data":"0e293c12be0bbd5530533ebca34fda223774381927a552096f4ece03f4a3f731"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.568012 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.567985159 podStartE2EDuration="2.567985159s" podCreationTimestamp="2025-10-11 03:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:24.512027582 +0000 UTC m=+152.070972367" watchObservedRunningTime="2025-10-11 03:08:24.567985159 +0000 UTC m=+152.126929944" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.575490 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed6df735-dd22-44b6-bcb0-c91199885a2f","Type":"ContainerStarted","Data":"2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9"} Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.577194 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.628845 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.955572 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:24 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:24 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:24 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.956005 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.963510 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:24 crc kubenswrapper[4754]: I1011 03:08:24.979380 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:08:25 crc kubenswrapper[4754]: W1011 03:08:25.047036 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode012e2a3_530e_4b61_85e8_16fcdd401ef5.slice/crio-ea4abb998f1635979b621a3c0143d1b297e683b393bff012fbedc22718b2ef80 WatchSource:0}: Error finding container ea4abb998f1635979b621a3c0143d1b297e683b393bff012fbedc22718b2ef80: Status 404 returned error can't find the container with id ea4abb998f1635979b621a3c0143d1b297e683b393bff012fbedc22718b2ef80 Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.065372 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume\") pod \"afd23974-0ea6-466b-a570-2d1d32c12a84\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.065420 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw825\" (UniqueName: \"kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825\") pod \"afd23974-0ea6-466b-a570-2d1d32c12a84\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.065465 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume\") pod \"afd23974-0ea6-466b-a570-2d1d32c12a84\" (UID: \"afd23974-0ea6-466b-a570-2d1d32c12a84\") " Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.066663 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume" (OuterVolumeSpecName: "config-volume") pod "afd23974-0ea6-466b-a570-2d1d32c12a84" (UID: "afd23974-0ea6-466b-a570-2d1d32c12a84"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.073208 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825" (OuterVolumeSpecName: "kube-api-access-bw825") pod "afd23974-0ea6-466b-a570-2d1d32c12a84" (UID: "afd23974-0ea6-466b-a570-2d1d32c12a84"). InnerVolumeSpecName "kube-api-access-bw825". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.075645 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "afd23974-0ea6-466b-a570-2d1d32c12a84" (UID: "afd23974-0ea6-466b-a570-2d1d32c12a84"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.169329 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afd23974-0ea6-466b-a570-2d1d32c12a84-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.169359 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw825\" (UniqueName: \"kubernetes.io/projected/afd23974-0ea6-466b-a570-2d1d32c12a84-kube-api-access-bw825\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.169369 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afd23974-0ea6-466b-a570-2d1d32c12a84-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.603706 4754 generic.go:334] "Generic (PLEG): container finished" podID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerID="9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966" exitCode=0 Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.603780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerDied","Data":"9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.604421 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerStarted","Data":"ce875a73958c9efae188fcbfd57610e6b561f1be98d8d243529f88b3606e2b80"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.609444 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" event={"ID":"afd23974-0ea6-466b-a570-2d1d32c12a84","Type":"ContainerDied","Data":"580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.609473 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580dcd7b22a4ad416a41bae7985e563f2b3aafa59fe081a5083404c1ad1727e6" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.609550 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.617808 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed6df735-dd22-44b6-bcb0-c91199885a2f","Type":"ContainerStarted","Data":"5113f90886a2c9f204f5bdc7727194e4ea22d6585f07b27cfa4929a11d0bfc7b"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.629323 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ddc52c0-5340-40e5-8a26-3b6188a72b70" containerID="b3a8f225c2464c449bed982166c16bef02416b17909fdc24bce1b5e8faa9d424" exitCode=0 Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.629404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ddc52c0-5340-40e5-8a26-3b6188a72b70","Type":"ContainerDied","Data":"b3a8f225c2464c449bed982166c16bef02416b17909fdc24bce1b5e8faa9d424"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.632615 4754 generic.go:334] "Generic (PLEG): container finished" podID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerID="6f8e874099eb798b8f4b365b393fd346fc376c66c1a801b4aaa98bf7ca9ab1b6" exitCode=0 Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.632662 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerDied","Data":"6f8e874099eb798b8f4b365b393fd346fc376c66c1a801b4aaa98bf7ca9ab1b6"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.632685 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerStarted","Data":"ea4abb998f1635979b621a3c0143d1b297e683b393bff012fbedc22718b2ef80"} Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.648917 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.648861991 podStartE2EDuration="2.648861991s" podCreationTimestamp="2025-10-11 03:08:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:25.639365872 +0000 UTC m=+153.198310667" watchObservedRunningTime="2025-10-11 03:08:25.648861991 +0000 UTC m=+153.207806776" Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.954484 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:25 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:25 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:25 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:25 crc kubenswrapper[4754]: I1011 03:08:25.954565 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:26 crc kubenswrapper[4754]: I1011 03:08:26.671073 4754 generic.go:334] "Generic (PLEG): container finished" podID="ed6df735-dd22-44b6-bcb0-c91199885a2f" containerID="5113f90886a2c9f204f5bdc7727194e4ea22d6585f07b27cfa4929a11d0bfc7b" exitCode=0 Oct 11 03:08:26 crc kubenswrapper[4754]: I1011 03:08:26.671182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed6df735-dd22-44b6-bcb0-c91199885a2f","Type":"ContainerDied","Data":"5113f90886a2c9f204f5bdc7727194e4ea22d6585f07b27cfa4929a11d0bfc7b"} Oct 11 03:08:26 crc kubenswrapper[4754]: I1011 03:08:26.953409 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:26 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:26 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:26 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:26 crc kubenswrapper[4754]: I1011 03:08:26.953822 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:26 crc kubenswrapper[4754]: I1011 03:08:26.977881 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.126607 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access\") pod \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.126766 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir\") pod \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\" (UID: \"4ddc52c0-5340-40e5-8a26-3b6188a72b70\") " Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.126899 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4ddc52c0-5340-40e5-8a26-3b6188a72b70" (UID: "4ddc52c0-5340-40e5-8a26-3b6188a72b70"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.127085 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.132565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4ddc52c0-5340-40e5-8a26-3b6188a72b70" (UID: "4ddc52c0-5340-40e5-8a26-3b6188a72b70"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.228677 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ddc52c0-5340-40e5-8a26-3b6188a72b70-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.694223 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.695044 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ddc52c0-5340-40e5-8a26-3b6188a72b70","Type":"ContainerDied","Data":"64e378df92b1f2ed33a21bee36438b1d77f6154da8fa342319b47c7da5790aae"} Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.695147 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64e378df92b1f2ed33a21bee36438b1d77f6154da8fa342319b47c7da5790aae" Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.951936 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:27 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:27 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:27 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:27 crc kubenswrapper[4754]: I1011 03:08:27.952072 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:28 crc kubenswrapper[4754]: I1011 03:08:28.157839 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:28 crc kubenswrapper[4754]: I1011 03:08:28.162989 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-42nc2" Oct 11 03:08:28 crc kubenswrapper[4754]: I1011 03:08:28.916300 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kl4ld" Oct 11 03:08:28 crc kubenswrapper[4754]: I1011 03:08:28.952843 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:28 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:28 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:28 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:28 crc kubenswrapper[4754]: I1011 03:08:28.952920 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:29 crc kubenswrapper[4754]: I1011 03:08:29.951784 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:29 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:29 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:29 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:29 crc kubenswrapper[4754]: I1011 03:08:29.951891 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:30 crc kubenswrapper[4754]: I1011 03:08:30.744706 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:08:30 crc kubenswrapper[4754]: I1011 03:08:30.745697 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:08:30 crc kubenswrapper[4754]: I1011 03:08:30.952244 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:30 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:30 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:30 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:30 crc kubenswrapper[4754]: I1011 03:08:30.952312 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:31 crc kubenswrapper[4754]: I1011 03:08:31.951774 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:31 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:31 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:31 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:31 crc kubenswrapper[4754]: I1011 03:08:31.951864 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:32 crc kubenswrapper[4754]: I1011 03:08:32.952455 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:32 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:32 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:32 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:32 crc kubenswrapper[4754]: I1011 03:08:32.952898 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:32 crc kubenswrapper[4754]: I1011 03:08:32.964078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-h69hd" Oct 11 03:08:33 crc kubenswrapper[4754]: I1011 03:08:33.145976 4754 patch_prober.go:28] interesting pod/console-f9d7485db-m25hk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 11 03:08:33 crc kubenswrapper[4754]: I1011 03:08:33.146048 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m25hk" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 11 03:08:33 crc kubenswrapper[4754]: I1011 03:08:33.951824 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:33 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:33 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:33 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:33 crc kubenswrapper[4754]: I1011 03:08:33.951898 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:34 crc kubenswrapper[4754]: I1011 03:08:34.952458 4754 patch_prober.go:28] interesting pod/router-default-5444994796-wq6lw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:08:34 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 11 03:08:34 crc kubenswrapper[4754]: [+]process-running ok Oct 11 03:08:34 crc kubenswrapper[4754]: healthz check failed Oct 11 03:08:34 crc kubenswrapper[4754]: I1011 03:08:34.952520 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wq6lw" podUID="7fc4ea3d-de59-4461-b58d-e74cac57dcf0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.365778 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.402537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access\") pod \"ed6df735-dd22-44b6-bcb0-c91199885a2f\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.402773 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir\") pod \"ed6df735-dd22-44b6-bcb0-c91199885a2f\" (UID: \"ed6df735-dd22-44b6-bcb0-c91199885a2f\") " Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.402949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed6df735-dd22-44b6-bcb0-c91199885a2f" (UID: "ed6df735-dd22-44b6-bcb0-c91199885a2f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.403070 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed6df735-dd22-44b6-bcb0-c91199885a2f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.411676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed6df735-dd22-44b6-bcb0-c91199885a2f" (UID: "ed6df735-dd22-44b6-bcb0-c91199885a2f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.503663 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed6df735-dd22-44b6-bcb0-c91199885a2f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.779181 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed6df735-dd22-44b6-bcb0-c91199885a2f","Type":"ContainerDied","Data":"2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9"} Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.779242 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ee2e04f95bf6f0d7ba2d436430a503d684dfa940d207af287203445f22922b9" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.779318 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.953156 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:35 crc kubenswrapper[4754]: I1011 03:08:35.956145 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-wq6lw" Oct 11 03:08:39 crc kubenswrapper[4754]: I1011 03:08:39.917763 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:39 crc kubenswrapper[4754]: I1011 03:08:39.925441 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5e7151-ac6f-4da4-9cc8-39639136e005-metrics-certs\") pod \"network-metrics-daemon-b6s82\" (UID: \"fb5e7151-ac6f-4da4-9cc8-39639136e005\") " pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:40 crc kubenswrapper[4754]: I1011 03:08:40.004446 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6s82" Oct 11 03:08:41 crc kubenswrapper[4754]: I1011 03:08:41.825800 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:08:43 crc kubenswrapper[4754]: I1011 03:08:43.148974 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:43 crc kubenswrapper[4754]: I1011 03:08:43.152879 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:08:47 crc kubenswrapper[4754]: E1011 03:08:47.054356 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 11 03:08:47 crc kubenswrapper[4754]: E1011 03:08:47.054864 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-87wgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bh57p_openshift-marketplace(fcf76dd9-1d7b-437b-9095-f5f043fce7f3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:08:47 crc kubenswrapper[4754]: E1011 03:08:47.056130 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-bh57p" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" Oct 11 03:08:48 crc kubenswrapper[4754]: E1011 03:08:48.359773 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bh57p" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" Oct 11 03:08:49 crc kubenswrapper[4754]: E1011 03:08:49.291045 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 11 03:08:49 crc kubenswrapper[4754]: E1011 03:08:49.291234 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q56ll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-z6p8f_openshift-marketplace(6a6ed923-5b2f-4966-971b-70bf271eb7a6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:08:49 crc kubenswrapper[4754]: E1011 03:08:49.292685 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-z6p8f" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" Oct 11 03:08:50 crc kubenswrapper[4754]: E1011 03:08:50.211310 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-z6p8f" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.523677 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.524800 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lwqtd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-r727p_openshift-marketplace(e1382c7c-c62d-46f2-ba3c-a325af801eaf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.526021 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-r727p" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.566617 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.566865 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crmg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gxp22_openshift-marketplace(7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.567863 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.568156 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kthrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8xmn8_openshift-marketplace(e012e2a3-530e-4b61-85e8-16fcdd401ef5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.569304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8xmn8" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.570264 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gxp22" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" Oct 11 03:08:53 crc kubenswrapper[4754]: I1011 03:08:53.880155 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nbpp4" Oct 11 03:08:53 crc kubenswrapper[4754]: I1011 03:08:53.904354 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerStarted","Data":"c1fcec861151f46f4f7565e519c9eaf4c4f48ba0645b739620c9c111365a544c"} Oct 11 03:08:53 crc kubenswrapper[4754]: I1011 03:08:53.906247 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerStarted","Data":"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949"} Oct 11 03:08:53 crc kubenswrapper[4754]: I1011 03:08:53.911954 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerStarted","Data":"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184"} Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.916210 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gxp22" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.916336 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-r727p" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" Oct 11 03:08:53 crc kubenswrapper[4754]: E1011 03:08:53.916404 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8xmn8" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" Oct 11 03:08:53 crc kubenswrapper[4754]: I1011 03:08:53.957000 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b6s82"] Oct 11 03:08:53 crc kubenswrapper[4754]: W1011 03:08:53.970511 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb5e7151_ac6f_4da4_9cc8_39639136e005.slice/crio-fb4887939bcf7436437bd9c1c36da65b9f96c6a3a0f2b1e6eb512a515490c849 WatchSource:0}: Error finding container fb4887939bcf7436437bd9c1c36da65b9f96c6a3a0f2b1e6eb512a515490c849: Status 404 returned error can't find the container with id fb4887939bcf7436437bd9c1c36da65b9f96c6a3a0f2b1e6eb512a515490c849 Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.920619 4754 generic.go:334] "Generic (PLEG): container finished" podID="49c5c869-51bd-442b-8567-98cfd45c4765" containerID="4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949" exitCode=0 Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.920721 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerDied","Data":"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.921216 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerStarted","Data":"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.924299 4754 generic.go:334] "Generic (PLEG): container finished" podID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerID="a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184" exitCode=0 Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.924371 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerDied","Data":"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.927802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6s82" event={"ID":"fb5e7151-ac6f-4da4-9cc8-39639136e005","Type":"ContainerStarted","Data":"c6d2390e5e6f22067849226ef0733136326b68a2ad60da1beb247e3a39a5a038"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.927849 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6s82" event={"ID":"fb5e7151-ac6f-4da4-9cc8-39639136e005","Type":"ContainerStarted","Data":"48d00520c3ec78ba1df7227d8cac87a66ad837f387a282a119520f331631e900"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.927863 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6s82" event={"ID":"fb5e7151-ac6f-4da4-9cc8-39639136e005","Type":"ContainerStarted","Data":"fb4887939bcf7436437bd9c1c36da65b9f96c6a3a0f2b1e6eb512a515490c849"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.929723 4754 generic.go:334] "Generic (PLEG): container finished" podID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerID="c1fcec861151f46f4f7565e519c9eaf4c4f48ba0645b739620c9c111365a544c" exitCode=0 Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.929758 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerDied","Data":"c1fcec861151f46f4f7565e519c9eaf4c4f48ba0645b739620c9c111365a544c"} Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.962771 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b6s82" podStartSLOduration=157.9627442 podStartE2EDuration="2m37.9627442s" podCreationTimestamp="2025-10-11 03:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:08:54.962399407 +0000 UTC m=+182.521344232" watchObservedRunningTime="2025-10-11 03:08:54.9627442 +0000 UTC m=+182.521688985" Oct 11 03:08:54 crc kubenswrapper[4754]: I1011 03:08:54.964748 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4sn62" podStartSLOduration=2.107940304 podStartE2EDuration="31.964734621s" podCreationTimestamp="2025-10-11 03:08:23 +0000 UTC" firstStartedPulling="2025-10-11 03:08:24.530116558 +0000 UTC m=+152.089061343" lastFinishedPulling="2025-10-11 03:08:54.386910875 +0000 UTC m=+181.945855660" observedRunningTime="2025-10-11 03:08:54.943341967 +0000 UTC m=+182.502286742" watchObservedRunningTime="2025-10-11 03:08:54.964734621 +0000 UTC m=+182.523679416" Oct 11 03:08:55 crc kubenswrapper[4754]: I1011 03:08:55.939306 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerStarted","Data":"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c"} Oct 11 03:08:55 crc kubenswrapper[4754]: I1011 03:08:55.943343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerStarted","Data":"a1e4b26385ffc0e39851dddadc085435a96e013b2874d0b1c030494659cc6e03"} Oct 11 03:08:56 crc kubenswrapper[4754]: I1011 03:08:56.003274 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7nb8q" podStartSLOduration=3.065567517 podStartE2EDuration="35.003249291s" podCreationTimestamp="2025-10-11 03:08:21 +0000 UTC" firstStartedPulling="2025-10-11 03:08:23.409432916 +0000 UTC m=+150.968377701" lastFinishedPulling="2025-10-11 03:08:55.34711469 +0000 UTC m=+182.906059475" observedRunningTime="2025-10-11 03:08:56.000385898 +0000 UTC m=+183.559330693" watchObservedRunningTime="2025-10-11 03:08:56.003249291 +0000 UTC m=+183.562194066" Oct 11 03:08:56 crc kubenswrapper[4754]: I1011 03:08:56.005636 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hvc2g" podStartSLOduration=3.2049203 podStartE2EDuration="33.005625995s" podCreationTimestamp="2025-10-11 03:08:23 +0000 UTC" firstStartedPulling="2025-10-11 03:08:25.61687952 +0000 UTC m=+153.175824305" lastFinishedPulling="2025-10-11 03:08:55.417585215 +0000 UTC m=+182.976530000" observedRunningTime="2025-10-11 03:08:55.976245687 +0000 UTC m=+183.535190472" watchObservedRunningTime="2025-10-11 03:08:56.005625995 +0000 UTC m=+183.564570780" Oct 11 03:09:00 crc kubenswrapper[4754]: I1011 03:09:00.035521 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:09:00 crc kubenswrapper[4754]: I1011 03:09:00.736198 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:09:00 crc kubenswrapper[4754]: I1011 03:09:00.736571 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:09:00 crc kubenswrapper[4754]: I1011 03:09:00.971236 4754 generic.go:334] "Generic (PLEG): container finished" podID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerID="b9006d277151480e49d80c8343f9d2ea3470ddbb0dba79ff05617066f8633524" exitCode=0 Oct 11 03:09:00 crc kubenswrapper[4754]: I1011 03:09:00.971273 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerDied","Data":"b9006d277151480e49d80c8343f9d2ea3470ddbb0dba79ff05617066f8633524"} Oct 11 03:09:01 crc kubenswrapper[4754]: I1011 03:09:01.674448 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:01 crc kubenswrapper[4754]: I1011 03:09:01.674629 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:01 crc kubenswrapper[4754]: I1011 03:09:01.837358 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:01 crc kubenswrapper[4754]: I1011 03:09:01.980686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerStarted","Data":"adebe17796e2b87dbdc227806f274a7236717215531975e09b4f47d93eb69b43"} Oct 11 03:09:01 crc kubenswrapper[4754]: I1011 03:09:01.984327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerStarted","Data":"cab5d27e861aeb5b13dd81d81489af0c20e0105b0223d469f7b1168283722546"} Oct 11 03:09:02 crc kubenswrapper[4754]: I1011 03:09:02.041581 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:02 crc kubenswrapper[4754]: I1011 03:09:02.044034 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bh57p" podStartSLOduration=3.677681771 podStartE2EDuration="42.044009553s" podCreationTimestamp="2025-10-11 03:08:20 +0000 UTC" firstStartedPulling="2025-10-11 03:08:23.338250005 +0000 UTC m=+150.897194790" lastFinishedPulling="2025-10-11 03:09:01.704577737 +0000 UTC m=+189.263522572" observedRunningTime="2025-10-11 03:09:02.043465264 +0000 UTC m=+189.602410049" watchObservedRunningTime="2025-10-11 03:09:02.044009553 +0000 UTC m=+189.602954378" Oct 11 03:09:02 crc kubenswrapper[4754]: I1011 03:09:02.992464 4754 generic.go:334] "Generic (PLEG): container finished" podID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerID="adebe17796e2b87dbdc227806f274a7236717215531975e09b4f47d93eb69b43" exitCode=0 Oct 11 03:09:02 crc kubenswrapper[4754]: I1011 03:09:02.992595 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerDied","Data":"adebe17796e2b87dbdc227806f274a7236717215531975e09b4f47d93eb69b43"} Oct 11 03:09:03 crc kubenswrapper[4754]: I1011 03:09:03.621947 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:03 crc kubenswrapper[4754]: I1011 03:09:03.622422 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:03 crc kubenswrapper[4754]: I1011 03:09:03.660713 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:03 crc kubenswrapper[4754]: I1011 03:09:03.689068 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.002572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerStarted","Data":"ac571fd4404f71c10987465879f21112d0a2b2ee6cc9bf22ab9c738d195d9d75"} Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.024725 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z6p8f" podStartSLOduration=3.962354273 podStartE2EDuration="44.024700713s" podCreationTimestamp="2025-10-11 03:08:20 +0000 UTC" firstStartedPulling="2025-10-11 03:08:23.339144127 +0000 UTC m=+150.898088912" lastFinishedPulling="2025-10-11 03:09:03.401490567 +0000 UTC m=+190.960435352" observedRunningTime="2025-10-11 03:09:04.021323372 +0000 UTC m=+191.580268157" watchObservedRunningTime="2025-10-11 03:09:04.024700713 +0000 UTC m=+191.583645498" Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.045520 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.270746 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.271259 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:09:04 crc kubenswrapper[4754]: I1011 03:09:04.321101 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:09:05 crc kubenswrapper[4754]: I1011 03:09:05.008852 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7nb8q" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="registry-server" containerID="cri-o://a1e4b26385ffc0e39851dddadc085435a96e013b2874d0b1c030494659cc6e03" gracePeriod=2 Oct 11 03:09:05 crc kubenswrapper[4754]: I1011 03:09:05.095106 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.020317 4754 generic.go:334] "Generic (PLEG): container finished" podID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerID="a1e4b26385ffc0e39851dddadc085435a96e013b2874d0b1c030494659cc6e03" exitCode=0 Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.020413 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerDied","Data":"a1e4b26385ffc0e39851dddadc085435a96e013b2874d0b1c030494659cc6e03"} Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.474415 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.526488 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities\") pod \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.526547 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trngh\" (UniqueName: \"kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh\") pod \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.526602 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content\") pod \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\" (UID: \"bd0c4ffb-6ffd-4948-a148-8a07a8828f24\") " Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.527223 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities" (OuterVolumeSpecName: "utilities") pod "bd0c4ffb-6ffd-4948-a148-8a07a8828f24" (UID: "bd0c4ffb-6ffd-4948-a148-8a07a8828f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.534343 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh" (OuterVolumeSpecName: "kube-api-access-trngh") pod "bd0c4ffb-6ffd-4948-a148-8a07a8828f24" (UID: "bd0c4ffb-6ffd-4948-a148-8a07a8828f24"). InnerVolumeSpecName "kube-api-access-trngh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.570692 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd0c4ffb-6ffd-4948-a148-8a07a8828f24" (UID: "bd0c4ffb-6ffd-4948-a148-8a07a8828f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.627850 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.627899 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:06 crc kubenswrapper[4754]: I1011 03:09:06.627920 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trngh\" (UniqueName: \"kubernetes.io/projected/bd0c4ffb-6ffd-4948-a148-8a07a8828f24-kube-api-access-trngh\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.027667 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nb8q" event={"ID":"bd0c4ffb-6ffd-4948-a148-8a07a8828f24","Type":"ContainerDied","Data":"0a99827a4fb635d1aa77474037a61f2822e60cde0925315534dfdd4e3184c552"} Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.027696 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nb8q" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.028041 4754 scope.go:117] "RemoveContainer" containerID="a1e4b26385ffc0e39851dddadc085435a96e013b2874d0b1c030494659cc6e03" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.061072 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.063645 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7nb8q"] Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.101546 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" path="/var/lib/kubelet/pods/bd0c4ffb-6ffd-4948-a148-8a07a8828f24/volumes" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.459695 4754 scope.go:117] "RemoveContainer" containerID="c1fcec861151f46f4f7565e519c9eaf4c4f48ba0645b739620c9c111365a544c" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.488191 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.489015 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4sn62" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="registry-server" containerID="cri-o://fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b" gracePeriod=2 Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.497681 4754 scope.go:117] "RemoveContainer" containerID="7b8b911db2ece96c2f17bd8a7ba81fed20a5e044fe4927c4d483b52a0a636900" Oct 11 03:09:07 crc kubenswrapper[4754]: I1011 03:09:07.943518 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.035505 4754 generic.go:334] "Generic (PLEG): container finished" podID="49c5c869-51bd-442b-8567-98cfd45c4765" containerID="fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b" exitCode=0 Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.035554 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerDied","Data":"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b"} Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.035579 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sn62" event={"ID":"49c5c869-51bd-442b-8567-98cfd45c4765","Type":"ContainerDied","Data":"10ca9ff453c639290bbaf753c666fdc9321ee8efb76deb9c250a3c42d39b10eb"} Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.035578 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sn62" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.035595 4754 scope.go:117] "RemoveContainer" containerID="fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.044563 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities\") pod \"49c5c869-51bd-442b-8567-98cfd45c4765\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.044632 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px6td\" (UniqueName: \"kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td\") pod \"49c5c869-51bd-442b-8567-98cfd45c4765\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.044655 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content\") pod \"49c5c869-51bd-442b-8567-98cfd45c4765\" (UID: \"49c5c869-51bd-442b-8567-98cfd45c4765\") " Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.045912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities" (OuterVolumeSpecName: "utilities") pod "49c5c869-51bd-442b-8567-98cfd45c4765" (UID: "49c5c869-51bd-442b-8567-98cfd45c4765"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.051008 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td" (OuterVolumeSpecName: "kube-api-access-px6td") pod "49c5c869-51bd-442b-8567-98cfd45c4765" (UID: "49c5c869-51bd-442b-8567-98cfd45c4765"). InnerVolumeSpecName "kube-api-access-px6td". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.067268 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49c5c869-51bd-442b-8567-98cfd45c4765" (UID: "49c5c869-51bd-442b-8567-98cfd45c4765"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.069597 4754 scope.go:117] "RemoveContainer" containerID="4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.091313 4754 scope.go:117] "RemoveContainer" containerID="996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.107026 4754 scope.go:117] "RemoveContainer" containerID="fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b" Oct 11 03:09:08 crc kubenswrapper[4754]: E1011 03:09:08.107449 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b\": container with ID starting with fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b not found: ID does not exist" containerID="fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.107488 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b"} err="failed to get container status \"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b\": rpc error: code = NotFound desc = could not find container \"fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b\": container with ID starting with fdaec3231df194bad113acf4d1a3ae90fed1dddfbd8628a8de8366b31fb0d14b not found: ID does not exist" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.107542 4754 scope.go:117] "RemoveContainer" containerID="4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949" Oct 11 03:09:08 crc kubenswrapper[4754]: E1011 03:09:08.107752 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949\": container with ID starting with 4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949 not found: ID does not exist" containerID="4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.107781 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949"} err="failed to get container status \"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949\": rpc error: code = NotFound desc = could not find container \"4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949\": container with ID starting with 4512ef0c3b83a07d892afa7d91918baf584e8dde0a823cff615dc3bacbb4c949 not found: ID does not exist" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.107800 4754 scope.go:117] "RemoveContainer" containerID="996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281" Oct 11 03:09:08 crc kubenswrapper[4754]: E1011 03:09:08.108096 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281\": container with ID starting with 996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281 not found: ID does not exist" containerID="996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.108123 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281"} err="failed to get container status \"996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281\": rpc error: code = NotFound desc = could not find container \"996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281\": container with ID starting with 996e90a428b082cc78d788fe0530b7b701caba22f4b42e5b21d92592a1d89281 not found: ID does not exist" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.145927 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.145978 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px6td\" (UniqueName: \"kubernetes.io/projected/49c5c869-51bd-442b-8567-98cfd45c4765-kube-api-access-px6td\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.145989 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49c5c869-51bd-442b-8567-98cfd45c4765-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.371164 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:09:08 crc kubenswrapper[4754]: I1011 03:09:08.373320 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sn62"] Oct 11 03:09:09 crc kubenswrapper[4754]: I1011 03:09:09.044865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerStarted","Data":"99c05c3898b45eac4fbaa089f9505b4f06203e79f64bc6d9636cfb76ff1587d9"} Oct 11 03:09:09 crc kubenswrapper[4754]: I1011 03:09:09.054134 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerID="e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b" exitCode=0 Oct 11 03:09:09 crc kubenswrapper[4754]: I1011 03:09:09.054167 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerDied","Data":"e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b"} Oct 11 03:09:09 crc kubenswrapper[4754]: I1011 03:09:09.092076 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" path="/var/lib/kubelet/pods/49c5c869-51bd-442b-8567-98cfd45c4765/volumes" Oct 11 03:09:10 crc kubenswrapper[4754]: I1011 03:09:10.067082 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerStarted","Data":"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00"} Oct 11 03:09:10 crc kubenswrapper[4754]: I1011 03:09:10.071227 4754 generic.go:334] "Generic (PLEG): container finished" podID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerID="99c05c3898b45eac4fbaa089f9505b4f06203e79f64bc6d9636cfb76ff1587d9" exitCode=0 Oct 11 03:09:10 crc kubenswrapper[4754]: I1011 03:09:10.071283 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerDied","Data":"99c05c3898b45eac4fbaa089f9505b4f06203e79f64bc6d9636cfb76ff1587d9"} Oct 11 03:09:10 crc kubenswrapper[4754]: I1011 03:09:10.101816 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r727p" podStartSLOduration=3.1788968459999998 podStartE2EDuration="48.101792754s" podCreationTimestamp="2025-10-11 03:08:22 +0000 UTC" firstStartedPulling="2025-10-11 03:08:24.535845952 +0000 UTC m=+152.094790737" lastFinishedPulling="2025-10-11 03:09:09.45874186 +0000 UTC m=+197.017686645" observedRunningTime="2025-10-11 03:09:10.099618786 +0000 UTC m=+197.658563581" watchObservedRunningTime="2025-10-11 03:09:10.101792754 +0000 UTC m=+197.660737539" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.079891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerStarted","Data":"0a49ae3b7c7bf1cb4dc829020dca9be8eecb94b037f6f8c05cee034a900ba04f"} Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.083492 4754 generic.go:334] "Generic (PLEG): container finished" podID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerID="887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868" exitCode=0 Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.103130 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8xmn8" podStartSLOduration=2.111705216 podStartE2EDuration="47.103104612s" podCreationTimestamp="2025-10-11 03:08:24 +0000 UTC" firstStartedPulling="2025-10-11 03:08:25.638320635 +0000 UTC m=+153.197265420" lastFinishedPulling="2025-10-11 03:09:10.629720031 +0000 UTC m=+198.188664816" observedRunningTime="2025-10-11 03:09:11.099860945 +0000 UTC m=+198.658805750" watchObservedRunningTime="2025-10-11 03:09:11.103104612 +0000 UTC m=+198.662049397" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.103765 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerDied","Data":"887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868"} Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.103824 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.103840 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.139873 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.276884 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.277783 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:09:11 crc kubenswrapper[4754]: I1011 03:09:11.331002 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:09:12 crc kubenswrapper[4754]: I1011 03:09:12.186778 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:09:12 crc kubenswrapper[4754]: I1011 03:09:12.209783 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:09:13 crc kubenswrapper[4754]: I1011 03:09:13.107320 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerStarted","Data":"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf"} Oct 11 03:09:13 crc kubenswrapper[4754]: I1011 03:09:13.190802 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:09:13 crc kubenswrapper[4754]: I1011 03:09:13.191238 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:09:13 crc kubenswrapper[4754]: I1011 03:09:13.248832 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:09:13 crc kubenswrapper[4754]: I1011 03:09:13.277672 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxp22" podStartSLOduration=4.135725744 podStartE2EDuration="52.277652031s" podCreationTimestamp="2025-10-11 03:08:21 +0000 UTC" firstStartedPulling="2025-10-11 03:08:23.379351782 +0000 UTC m=+150.938296567" lastFinishedPulling="2025-10-11 03:09:11.521278069 +0000 UTC m=+199.080222854" observedRunningTime="2025-10-11 03:09:13.130428593 +0000 UTC m=+200.689373448" watchObservedRunningTime="2025-10-11 03:09:13.277652031 +0000 UTC m=+200.836596816" Oct 11 03:09:14 crc kubenswrapper[4754]: I1011 03:09:14.166344 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:09:14 crc kubenswrapper[4754]: I1011 03:09:14.629680 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:14 crc kubenswrapper[4754]: I1011 03:09:14.630271 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:15 crc kubenswrapper[4754]: I1011 03:09:15.704045 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8xmn8" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="registry-server" probeResult="failure" output=< Oct 11 03:09:15 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 11 03:09:15 crc kubenswrapper[4754]: > Oct 11 03:09:19 crc kubenswrapper[4754]: I1011 03:09:19.666602 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:09:21 crc kubenswrapper[4754]: I1011 03:09:21.464197 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:21 crc kubenswrapper[4754]: I1011 03:09:21.464667 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:21 crc kubenswrapper[4754]: I1011 03:09:21.503919 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:22 crc kubenswrapper[4754]: I1011 03:09:22.219551 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:24 crc kubenswrapper[4754]: I1011 03:09:24.482332 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:09:24 crc kubenswrapper[4754]: I1011 03:09:24.483053 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gxp22" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="registry-server" containerID="cri-o://0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf" gracePeriod=2 Oct 11 03:09:24 crc kubenswrapper[4754]: I1011 03:09:24.695703 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:24 crc kubenswrapper[4754]: I1011 03:09:24.799352 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:24 crc kubenswrapper[4754]: I1011 03:09:24.975419 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.113678 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities\") pod \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.113912 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crmg5\" (UniqueName: \"kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5\") pod \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.113952 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content\") pod \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\" (UID: \"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b\") " Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.114594 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities" (OuterVolumeSpecName: "utilities") pod "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" (UID: "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.120917 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5" (OuterVolumeSpecName: "kube-api-access-crmg5") pod "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" (UID: "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b"). InnerVolumeSpecName "kube-api-access-crmg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.170698 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" (UID: "7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.179906 4754 generic.go:334] "Generic (PLEG): container finished" podID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerID="0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf" exitCode=0 Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.180066 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxp22" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.180131 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerDied","Data":"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf"} Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.180190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxp22" event={"ID":"7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b","Type":"ContainerDied","Data":"c0a85b75423dffe538cbddc702bc2b07443e5d60e4f4e8c55620045c77d82bca"} Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.180213 4754 scope.go:117] "RemoveContainer" containerID="0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.206932 4754 scope.go:117] "RemoveContainer" containerID="887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.213077 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.215642 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crmg5\" (UniqueName: \"kubernetes.io/projected/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-kube-api-access-crmg5\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.215676 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.215695 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.215779 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gxp22"] Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.241366 4754 scope.go:117] "RemoveContainer" containerID="be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.257088 4754 scope.go:117] "RemoveContainer" containerID="0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf" Oct 11 03:09:25 crc kubenswrapper[4754]: E1011 03:09:25.257630 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf\": container with ID starting with 0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf not found: ID does not exist" containerID="0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.257674 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf"} err="failed to get container status \"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf\": rpc error: code = NotFound desc = could not find container \"0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf\": container with ID starting with 0ab0e79c3330ec963f7185cfe29ea77a334c2e2349c1e2f4105e1259be5d1ecf not found: ID does not exist" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.257712 4754 scope.go:117] "RemoveContainer" containerID="887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868" Oct 11 03:09:25 crc kubenswrapper[4754]: E1011 03:09:25.258111 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868\": container with ID starting with 887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868 not found: ID does not exist" containerID="887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.258135 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868"} err="failed to get container status \"887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868\": rpc error: code = NotFound desc = could not find container \"887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868\": container with ID starting with 887feeb5c6d92c47867cc6508f01a8be3d8997f2fe5c5388e68ce18ff9e15868 not found: ID does not exist" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.258152 4754 scope.go:117] "RemoveContainer" containerID="be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9" Oct 11 03:09:25 crc kubenswrapper[4754]: E1011 03:09:25.258404 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9\": container with ID starting with be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9 not found: ID does not exist" containerID="be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9" Oct 11 03:09:25 crc kubenswrapper[4754]: I1011 03:09:25.258432 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9"} err="failed to get container status \"be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9\": rpc error: code = NotFound desc = could not find container \"be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9\": container with ID starting with be9feab76d187fa207d6ff0de34b861817211bb9e85a9c0325db076de3f308b9 not found: ID does not exist" Oct 11 03:09:27 crc kubenswrapper[4754]: I1011 03:09:27.093596 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" path="/var/lib/kubelet/pods/7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b/volumes" Oct 11 03:09:27 crc kubenswrapper[4754]: I1011 03:09:27.883872 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:09:27 crc kubenswrapper[4754]: I1011 03:09:27.884245 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8xmn8" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="registry-server" containerID="cri-o://0a49ae3b7c7bf1cb4dc829020dca9be8eecb94b037f6f8c05cee034a900ba04f" gracePeriod=2 Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.202800 4754 generic.go:334] "Generic (PLEG): container finished" podID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerID="0a49ae3b7c7bf1cb4dc829020dca9be8eecb94b037f6f8c05cee034a900ba04f" exitCode=0 Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.202860 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerDied","Data":"0a49ae3b7c7bf1cb4dc829020dca9be8eecb94b037f6f8c05cee034a900ba04f"} Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.296689 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.365371 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities" (OuterVolumeSpecName: "utilities") pod "e012e2a3-530e-4b61-85e8-16fcdd401ef5" (UID: "e012e2a3-530e-4b61-85e8-16fcdd401ef5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.365425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities\") pod \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.365529 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content\") pod \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.365558 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kthrc\" (UniqueName: \"kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc\") pod \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\" (UID: \"e012e2a3-530e-4b61-85e8-16fcdd401ef5\") " Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.367736 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.371768 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc" (OuterVolumeSpecName: "kube-api-access-kthrc") pod "e012e2a3-530e-4b61-85e8-16fcdd401ef5" (UID: "e012e2a3-530e-4b61-85e8-16fcdd401ef5"). InnerVolumeSpecName "kube-api-access-kthrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.469495 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kthrc\" (UniqueName: \"kubernetes.io/projected/e012e2a3-530e-4b61-85e8-16fcdd401ef5-kube-api-access-kthrc\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.483445 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e012e2a3-530e-4b61-85e8-16fcdd401ef5" (UID: "e012e2a3-530e-4b61-85e8-16fcdd401ef5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:28 crc kubenswrapper[4754]: I1011 03:09:28.571032 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e012e2a3-530e-4b61-85e8-16fcdd401ef5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.210819 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xmn8" event={"ID":"e012e2a3-530e-4b61-85e8-16fcdd401ef5","Type":"ContainerDied","Data":"ea4abb998f1635979b621a3c0143d1b297e683b393bff012fbedc22718b2ef80"} Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.210889 4754 scope.go:117] "RemoveContainer" containerID="0a49ae3b7c7bf1cb4dc829020dca9be8eecb94b037f6f8c05cee034a900ba04f" Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.211012 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xmn8" Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.234656 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.248235 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8xmn8"] Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.248241 4754 scope.go:117] "RemoveContainer" containerID="99c05c3898b45eac4fbaa089f9505b4f06203e79f64bc6d9636cfb76ff1587d9" Oct 11 03:09:29 crc kubenswrapper[4754]: I1011 03:09:29.262895 4754 scope.go:117] "RemoveContainer" containerID="6f8e874099eb798b8f4b365b393fd346fc376c66c1a801b4aaa98bf7ca9ab1b6" Oct 11 03:09:30 crc kubenswrapper[4754]: I1011 03:09:30.737261 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:09:30 crc kubenswrapper[4754]: I1011 03:09:30.737341 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:09:30 crc kubenswrapper[4754]: I1011 03:09:30.737407 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:09:30 crc kubenswrapper[4754]: I1011 03:09:30.738022 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:09:30 crc kubenswrapper[4754]: I1011 03:09:30.738406 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195" gracePeriod=600 Oct 11 03:09:31 crc kubenswrapper[4754]: I1011 03:09:31.091247 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" path="/var/lib/kubelet/pods/e012e2a3-530e-4b61-85e8-16fcdd401ef5/volumes" Oct 11 03:09:31 crc kubenswrapper[4754]: I1011 03:09:31.227332 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195" exitCode=0 Oct 11 03:09:31 crc kubenswrapper[4754]: I1011 03:09:31.227386 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195"} Oct 11 03:09:31 crc kubenswrapper[4754]: I1011 03:09:31.227422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd"} Oct 11 03:09:44 crc kubenswrapper[4754]: I1011 03:09:44.699252 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerName="oauth-openshift" containerID="cri-o://f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552" gracePeriod=15 Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.188506 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.243684 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6499b46898-qwfkw"] Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.243988 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244006 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244020 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244028 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244041 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244050 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244059 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244069 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244080 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddc52c0-5340-40e5-8a26-3b6188a72b70" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244089 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddc52c0-5340-40e5-8a26-3b6188a72b70" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244101 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244109 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244122 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd23974-0ea6-466b-a570-2d1d32c12a84" containerName="collect-profiles" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244130 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd23974-0ea6-466b-a570-2d1d32c12a84" containerName="collect-profiles" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244141 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244151 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244162 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerName="oauth-openshift" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244170 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerName="oauth-openshift" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244180 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6df735-dd22-44b6-bcb0-c91199885a2f" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244188 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6df735-dd22-44b6-bcb0-c91199885a2f" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244201 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244210 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244222 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244230 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244244 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244253 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244264 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244275 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="extract-utilities" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244292 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244302 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="extract-content" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.244315 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244329 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244457 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddc52c0-5340-40e5-8a26-3b6188a72b70" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244473 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6df735-dd22-44b6-bcb0-c91199885a2f" containerName="pruner" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244488 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c5c869-51bd-442b-8567-98cfd45c4765" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244500 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerName="oauth-openshift" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244509 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd0c4ffb-6ffd-4948-a148-8a07a8828f24" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244520 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e012e2a3-530e-4b61-85e8-16fcdd401ef5" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244530 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bfd6246-2abb-4fd6-9bab-9d0befa5ad7b" containerName="registry-server" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.244542 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd23974-0ea6-466b-a570-2d1d32c12a84" containerName="collect-profiles" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.245050 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.259161 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6499b46898-qwfkw"] Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.311368 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.312461 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.312670 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.312826 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.312868 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.312940 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313029 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sr66\" (UniqueName: \"kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313165 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313203 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313270 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig\") pod \"01727067-3e38-424a-9adc-f7ffbe5e32e7\" (UID: \"01727067-3e38-424a-9adc-f7ffbe5e32e7\") " Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313391 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313407 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313457 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313643 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-policies\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313684 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313734 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-login\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-router-certs\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-service-ca\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313896 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-session\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313921 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.313957 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-error\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-dir\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314112 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njcjl\" (UniqueName: \"kubernetes.io/projected/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-kube-api-access-njcjl\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314147 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314177 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314220 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314290 4754 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314306 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.314321 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.315223 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.315944 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.318002 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.322939 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.328532 4754 generic.go:334] "Generic (PLEG): container finished" podID="01727067-3e38-424a-9adc-f7ffbe5e32e7" containerID="f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552" exitCode=0 Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.328914 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" event={"ID":"01727067-3e38-424a-9adc-f7ffbe5e32e7","Type":"ContainerDied","Data":"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552"} Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.329282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" event={"ID":"01727067-3e38-424a-9adc-f7ffbe5e32e7","Type":"ContainerDied","Data":"feca07a891a5ba1cf9ffac62de5f3f9c494ca4dac8dee3e5ec513d862a4aa385"} Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.329436 4754 scope.go:117] "RemoveContainer" containerID="f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.329916 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v89kr" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.337896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.338192 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66" (OuterVolumeSpecName: "kube-api-access-2sr66") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "kube-api-access-2sr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.338672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.338905 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.339866 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.340539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.340619 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "01727067-3e38-424a-9adc-f7ffbe5e32e7" (UID: "01727067-3e38-424a-9adc-f7ffbe5e32e7"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.378427 4754 scope.go:117] "RemoveContainer" containerID="f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552" Oct 11 03:09:45 crc kubenswrapper[4754]: E1011 03:09:45.380639 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552\": container with ID starting with f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552 not found: ID does not exist" containerID="f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.380737 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552"} err="failed to get container status \"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552\": rpc error: code = NotFound desc = could not find container \"f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552\": container with ID starting with f65ae3544981402859de57c961947776c1242c886c7e624e3d64ad845b14c552 not found: ID does not exist" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-session\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416234 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416259 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416312 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-error\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-dir\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416361 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njcjl\" (UniqueName: \"kubernetes.io/projected/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-kube-api-access-njcjl\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416386 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416404 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-policies\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416487 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-login\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416530 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-router-certs\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416557 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-service-ca\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416613 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416626 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416637 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sr66\" (UniqueName: \"kubernetes.io/projected/01727067-3e38-424a-9adc-f7ffbe5e32e7-kube-api-access-2sr66\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416647 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416657 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416668 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416678 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416687 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416696 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416705 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416715 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/01727067-3e38-424a-9adc-f7ffbe5e32e7-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.416765 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-dir\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.418088 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-service-ca\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.418110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-audit-policies\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.418192 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.420154 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.420602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-login\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.420751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.421136 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-error\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.421326 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.421483 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-router-certs\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.421763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.422358 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-session\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.422785 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.437468 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njcjl\" (UniqueName: \"kubernetes.io/projected/c612d9f9-b481-43e8-9c92-eb9b1b3fb343-kube-api-access-njcjl\") pod \"oauth-openshift-6499b46898-qwfkw\" (UID: \"c612d9f9-b481-43e8-9c92-eb9b1b3fb343\") " pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.588385 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.690142 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:09:45 crc kubenswrapper[4754]: I1011 03:09:45.693753 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v89kr"] Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.003852 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6499b46898-qwfkw"] Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.338750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" event={"ID":"c612d9f9-b481-43e8-9c92-eb9b1b3fb343","Type":"ContainerStarted","Data":"ffd6815c17685201cd2fd111814d5f79cc5a7b882827f642976e6568f992eac4"} Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.339112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" event={"ID":"c612d9f9-b481-43e8-9c92-eb9b1b3fb343","Type":"ContainerStarted","Data":"1daf0ede6e2d2fc501e762c01c103279f0605e60c4cd586ef62a1ae77f83d8bf"} Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.340150 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.340925 4754 patch_prober.go:28] interesting pod/oauth-openshift-6499b46898-qwfkw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" start-of-body= Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.340989 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" podUID="c612d9f9-b481-43e8-9c92-eb9b1b3fb343" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" Oct 11 03:09:46 crc kubenswrapper[4754]: I1011 03:09:46.376014 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" podStartSLOduration=27.375989401 podStartE2EDuration="27.375989401s" podCreationTimestamp="2025-10-11 03:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:09:46.36657677 +0000 UTC m=+233.925521565" watchObservedRunningTime="2025-10-11 03:09:46.375989401 +0000 UTC m=+233.934934196" Oct 11 03:09:47 crc kubenswrapper[4754]: I1011 03:09:47.096733 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01727067-3e38-424a-9adc-f7ffbe5e32e7" path="/var/lib/kubelet/pods/01727067-3e38-424a-9adc-f7ffbe5e32e7/volumes" Oct 11 03:09:47 crc kubenswrapper[4754]: I1011 03:09:47.359338 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6499b46898-qwfkw" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.245254 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.246150 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bh57p" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="registry-server" containerID="cri-o://cab5d27e861aeb5b13dd81d81489af0c20e0105b0223d469f7b1168283722546" gracePeriod=30 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.275175 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.277288 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z6p8f" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="registry-server" containerID="cri-o://ac571fd4404f71c10987465879f21112d0a2b2ee6cc9bf22ab9c738d195d9d75" gracePeriod=30 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.294348 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.295034 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" containerID="cri-o://3b7e29745a134eb2afb18b252e726099e6e9a8d1f4c650ea8059de67cacc1e2d" gracePeriod=30 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.306465 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.306737 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r727p" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="registry-server" containerID="cri-o://3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00" gracePeriod=30 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.308379 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.308639 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hvc2g" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="registry-server" containerID="cri-o://2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c" gracePeriod=30 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.330565 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46rsl"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.331457 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.336346 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46rsl"] Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.417808 4754 generic.go:334] "Generic (PLEG): container finished" podID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerID="3b7e29745a134eb2afb18b252e726099e6e9a8d1f4c650ea8059de67cacc1e2d" exitCode=0 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.417856 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" event={"ID":"d77034e9-1126-4903-8484-a99e47d2aa7f","Type":"ContainerDied","Data":"3b7e29745a134eb2afb18b252e726099e6e9a8d1f4c650ea8059de67cacc1e2d"} Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.423157 4754 generic.go:334] "Generic (PLEG): container finished" podID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerID="cab5d27e861aeb5b13dd81d81489af0c20e0105b0223d469f7b1168283722546" exitCode=0 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.423243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerDied","Data":"cab5d27e861aeb5b13dd81d81489af0c20e0105b0223d469f7b1168283722546"} Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.423709 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.423771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v8cs\" (UniqueName: \"kubernetes.io/projected/ed58c00a-acd3-4c68-847f-2c731a66b7bf-kube-api-access-9v8cs\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.423846 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.425737 4754 generic.go:334] "Generic (PLEG): container finished" podID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerID="ac571fd4404f71c10987465879f21112d0a2b2ee6cc9bf22ab9c738d195d9d75" exitCode=0 Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.425776 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerDied","Data":"ac571fd4404f71c10987465879f21112d0a2b2ee6cc9bf22ab9c738d195d9d75"} Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.524708 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.524774 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v8cs\" (UniqueName: \"kubernetes.io/projected/ed58c00a-acd3-4c68-847f-2c731a66b7bf-kube-api-access-9v8cs\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.524812 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.525981 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.539846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed58c00a-acd3-4c68-847f-2c731a66b7bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.545249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v8cs\" (UniqueName: \"kubernetes.io/projected/ed58c00a-acd3-4c68-847f-2c731a66b7bf-kube-api-access-9v8cs\") pod \"marketplace-operator-79b997595-46rsl\" (UID: \"ed58c00a-acd3-4c68-847f-2c731a66b7bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.719338 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.739901 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.756387 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.761332 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.779022 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.832886 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities\") pod \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.832947 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content\") pod \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.832988 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca\") pod \"d77034e9-1126-4903-8484-a99e47d2aa7f\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.834030 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d77034e9-1126-4903-8484-a99e47d2aa7f" (UID: "d77034e9-1126-4903-8484-a99e47d2aa7f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.834152 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities" (OuterVolumeSpecName: "utilities") pod "e1382c7c-c62d-46f2-ba3c-a325af801eaf" (UID: "e1382c7c-c62d-46f2-ba3c-a325af801eaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.834221 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87wgk\" (UniqueName: \"kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk\") pod \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.834993 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwqtd\" (UniqueName: \"kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd\") pod \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\" (UID: \"e1382c7c-c62d-46f2-ba3c-a325af801eaf\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835061 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities\") pod \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835091 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content\") pod \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\" (UID: \"fcf76dd9-1d7b-437b-9095-f5f043fce7f3\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835168 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities\") pod \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8r52\" (UniqueName: \"kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52\") pod \"d77034e9-1126-4903-8484-a99e47d2aa7f\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835235 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q56ll\" (UniqueName: \"kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll\") pod \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835268 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics\") pod \"d77034e9-1126-4903-8484-a99e47d2aa7f\" (UID: \"d77034e9-1126-4903-8484-a99e47d2aa7f\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835303 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content\") pod \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\" (UID: \"6a6ed923-5b2f-4966-971b-70bf271eb7a6\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835625 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.835638 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.839528 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities" (OuterVolumeSpecName: "utilities") pod "fcf76dd9-1d7b-437b-9095-f5f043fce7f3" (UID: "fcf76dd9-1d7b-437b-9095-f5f043fce7f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.843579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities" (OuterVolumeSpecName: "utilities") pod "6a6ed923-5b2f-4966-971b-70bf271eb7a6" (UID: "6a6ed923-5b2f-4966-971b-70bf271eb7a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.850518 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk" (OuterVolumeSpecName: "kube-api-access-87wgk") pod "fcf76dd9-1d7b-437b-9095-f5f043fce7f3" (UID: "fcf76dd9-1d7b-437b-9095-f5f043fce7f3"). InnerVolumeSpecName "kube-api-access-87wgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.851159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52" (OuterVolumeSpecName: "kube-api-access-f8r52") pod "d77034e9-1126-4903-8484-a99e47d2aa7f" (UID: "d77034e9-1126-4903-8484-a99e47d2aa7f"). InnerVolumeSpecName "kube-api-access-f8r52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.851714 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d77034e9-1126-4903-8484-a99e47d2aa7f" (UID: "d77034e9-1126-4903-8484-a99e47d2aa7f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.855690 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd" (OuterVolumeSpecName: "kube-api-access-lwqtd") pod "e1382c7c-c62d-46f2-ba3c-a325af801eaf" (UID: "e1382c7c-c62d-46f2-ba3c-a325af801eaf"). InnerVolumeSpecName "kube-api-access-lwqtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.857670 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll" (OuterVolumeSpecName: "kube-api-access-q56ll") pod "6a6ed923-5b2f-4966-971b-70bf271eb7a6" (UID: "6a6ed923-5b2f-4966-971b-70bf271eb7a6"). InnerVolumeSpecName "kube-api-access-q56ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.896221 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.908784 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1382c7c-c62d-46f2-ba3c-a325af801eaf" (UID: "e1382c7c-c62d-46f2-ba3c-a325af801eaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.937016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcf76dd9-1d7b-437b-9095-f5f043fce7f3" (UID: "fcf76dd9-1d7b-437b-9095-f5f043fce7f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.937563 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities\") pod \"b499bb38-4c39-4faf-9606-0ac50fe5c588\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.937732 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content\") pod \"b499bb38-4c39-4faf-9606-0ac50fe5c588\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.937838 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gxp7\" (UniqueName: \"kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7\") pod \"b499bb38-4c39-4faf-9606-0ac50fe5c588\" (UID: \"b499bb38-4c39-4faf-9606-0ac50fe5c588\") " Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938149 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q56ll\" (UniqueName: \"kubernetes.io/projected/6a6ed923-5b2f-4966-971b-70bf271eb7a6-kube-api-access-q56ll\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938165 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d77034e9-1126-4903-8484-a99e47d2aa7f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938179 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1382c7c-c62d-46f2-ba3c-a325af801eaf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938191 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87wgk\" (UniqueName: \"kubernetes.io/projected/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-kube-api-access-87wgk\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938204 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwqtd\" (UniqueName: \"kubernetes.io/projected/e1382c7c-c62d-46f2-ba3c-a325af801eaf-kube-api-access-lwqtd\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938217 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938226 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf76dd9-1d7b-437b-9095-f5f043fce7f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938238 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.938246 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8r52\" (UniqueName: \"kubernetes.io/projected/d77034e9-1126-4903-8484-a99e47d2aa7f-kube-api-access-f8r52\") on node \"crc\" DevicePath \"\"" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.940129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities" (OuterVolumeSpecName: "utilities") pod "b499bb38-4c39-4faf-9606-0ac50fe5c588" (UID: "b499bb38-4c39-4faf-9606-0ac50fe5c588"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.942337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7" (OuterVolumeSpecName: "kube-api-access-2gxp7") pod "b499bb38-4c39-4faf-9606-0ac50fe5c588" (UID: "b499bb38-4c39-4faf-9606-0ac50fe5c588"). InnerVolumeSpecName "kube-api-access-2gxp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:09:59 crc kubenswrapper[4754]: I1011 03:09:59.978509 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a6ed923-5b2f-4966-971b-70bf271eb7a6" (UID: "6a6ed923-5b2f-4966-971b-70bf271eb7a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.039953 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gxp7\" (UniqueName: \"kubernetes.io/projected/b499bb38-4c39-4faf-9606-0ac50fe5c588-kube-api-access-2gxp7\") on node \"crc\" DevicePath \"\"" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.040024 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.040040 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a6ed923-5b2f-4966-971b-70bf271eb7a6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.072799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b499bb38-4c39-4faf-9606-0ac50fe5c588" (UID: "b499bb38-4c39-4faf-9606-0ac50fe5c588"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.141680 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b499bb38-4c39-4faf-9606-0ac50fe5c588-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.231112 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46rsl"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.446299 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6p8f" event={"ID":"6a6ed923-5b2f-4966-971b-70bf271eb7a6","Type":"ContainerDied","Data":"5c9cc2c32991ff5941c4787064be4df6007c92d407578159a0b891fb3e02fb7b"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.446782 4754 scope.go:117] "RemoveContainer" containerID="ac571fd4404f71c10987465879f21112d0a2b2ee6cc9bf22ab9c738d195d9d75" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.446597 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6p8f" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.451664 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" event={"ID":"ed58c00a-acd3-4c68-847f-2c731a66b7bf","Type":"ContainerStarted","Data":"acf9620676bd0228e13aa99220ab0feff73940edc39ac3635ddb627bbb23c7ce"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.451733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" event={"ID":"ed58c00a-acd3-4c68-847f-2c731a66b7bf","Type":"ContainerStarted","Data":"86ce14f69dcc2e2f5ebebe70616666d9ca80b0f525c6b6d73be1671087a1d7d3"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.451860 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.455154 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-46rsl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.455205 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" podUID="ed58c00a-acd3-4c68-847f-2c731a66b7bf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.455704 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.455842 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w5w2q" event={"ID":"d77034e9-1126-4903-8484-a99e47d2aa7f","Type":"ContainerDied","Data":"3f5e6b94f3a4708906952f5f7058d22b8df255d4a3afe7e230dab175fa64ae9b"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.458400 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh57p" event={"ID":"fcf76dd9-1d7b-437b-9095-f5f043fce7f3","Type":"ContainerDied","Data":"065ee6454e2701abdb433efbd333fa84c8620a4e63fbb36c93c4aed18a7daeda"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.458528 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh57p" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.466282 4754 scope.go:117] "RemoveContainer" containerID="adebe17796e2b87dbdc227806f274a7236717215531975e09b4f47d93eb69b43" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.470530 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" podStartSLOduration=1.470503074 podStartE2EDuration="1.470503074s" podCreationTimestamp="2025-10-11 03:09:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:10:00.466216863 +0000 UTC m=+248.025161648" watchObservedRunningTime="2025-10-11 03:10:00.470503074 +0000 UTC m=+248.029447859" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.475687 4754 generic.go:334] "Generic (PLEG): container finished" podID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerID="2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c" exitCode=0 Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.475785 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerDied","Data":"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.475791 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvc2g" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.475825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvc2g" event={"ID":"b499bb38-4c39-4faf-9606-0ac50fe5c588","Type":"ContainerDied","Data":"ce875a73958c9efae188fcbfd57610e6b561f1be98d8d243529f88b3606e2b80"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.480388 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerID="3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00" exitCode=0 Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.480437 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r727p" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.480459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerDied","Data":"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.480508 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r727p" event={"ID":"e1382c7c-c62d-46f2-ba3c-a325af801eaf","Type":"ContainerDied","Data":"0e293c12be0bbd5530533ebca34fda223774381927a552096f4ece03f4a3f731"} Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.492796 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.500505 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z6p8f"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.504606 4754 scope.go:117] "RemoveContainer" containerID="73a69642a26637111b462b6c4bcb90a12be59d70d62e31ec1d579c4e6f66ee3e" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.519796 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.523153 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w5w2q"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.530358 4754 scope.go:117] "RemoveContainer" containerID="3b7e29745a134eb2afb18b252e726099e6e9a8d1f4c650ea8059de67cacc1e2d" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.535904 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.547051 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bh57p"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.551087 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.554717 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r727p"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.557178 4754 scope.go:117] "RemoveContainer" containerID="cab5d27e861aeb5b13dd81d81489af0c20e0105b0223d469f7b1168283722546" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.563767 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.569600 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hvc2g"] Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.577230 4754 scope.go:117] "RemoveContainer" containerID="b9006d277151480e49d80c8343f9d2ea3470ddbb0dba79ff05617066f8633524" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.593270 4754 scope.go:117] "RemoveContainer" containerID="6ba6b12a8f6d2d3e2af75dee471a100e55c9d587b02ba765030d49685a95d3a7" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.610670 4754 scope.go:117] "RemoveContainer" containerID="2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.632054 4754 scope.go:117] "RemoveContainer" containerID="a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.651270 4754 scope.go:117] "RemoveContainer" containerID="9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.665889 4754 scope.go:117] "RemoveContainer" containerID="2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.666853 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c\": container with ID starting with 2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c not found: ID does not exist" containerID="2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.666981 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c"} err="failed to get container status \"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c\": rpc error: code = NotFound desc = could not find container \"2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c\": container with ID starting with 2487dc5d72778728fd1d1a94a6c4cb6d6ca510175db4846b83d7ce6cd14dd93c not found: ID does not exist" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.667095 4754 scope.go:117] "RemoveContainer" containerID="a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.667701 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184\": container with ID starting with a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184 not found: ID does not exist" containerID="a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.667743 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184"} err="failed to get container status \"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184\": rpc error: code = NotFound desc = could not find container \"a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184\": container with ID starting with a4954e666f65224df4a62c7304e32afb3e55936280292b68462b4e3e377d0184 not found: ID does not exist" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.667776 4754 scope.go:117] "RemoveContainer" containerID="9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.668067 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966\": container with ID starting with 9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966 not found: ID does not exist" containerID="9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.668162 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966"} err="failed to get container status \"9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966\": rpc error: code = NotFound desc = could not find container \"9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966\": container with ID starting with 9003050b98aba420f21806a27f69d1f28d11c8cffed582ba36709a0163b98966 not found: ID does not exist" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.668233 4754 scope.go:117] "RemoveContainer" containerID="3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.683539 4754 scope.go:117] "RemoveContainer" containerID="e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.699525 4754 scope.go:117] "RemoveContainer" containerID="027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.717680 4754 scope.go:117] "RemoveContainer" containerID="3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.718389 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00\": container with ID starting with 3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00 not found: ID does not exist" containerID="3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.718445 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00"} err="failed to get container status \"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00\": rpc error: code = NotFound desc = could not find container \"3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00\": container with ID starting with 3dbeba0d739302dddc0601b342409daedc7dd006b498e9e2aea53f684cb7ec00 not found: ID does not exist" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.718486 4754 scope.go:117] "RemoveContainer" containerID="e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.719383 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b\": container with ID starting with e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b not found: ID does not exist" containerID="e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.719511 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b"} err="failed to get container status \"e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b\": rpc error: code = NotFound desc = could not find container \"e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b\": container with ID starting with e15e0971428e8822cd86dd0aaf6dd903d4d54f479f7d48a10f49eeb9fb92969b not found: ID does not exist" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.719606 4754 scope.go:117] "RemoveContainer" containerID="027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144" Oct 11 03:10:00 crc kubenswrapper[4754]: E1011 03:10:00.720123 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144\": container with ID starting with 027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144 not found: ID does not exist" containerID="027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144" Oct 11 03:10:00 crc kubenswrapper[4754]: I1011 03:10:00.720149 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144"} err="failed to get container status \"027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144\": rpc error: code = NotFound desc = could not find container \"027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144\": container with ID starting with 027de4efeaa5eaa763e9b1b885b798bc0a3c9e604510a6acbaba3bde42d04144 not found: ID does not exist" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.090312 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" path="/var/lib/kubelet/pods/6a6ed923-5b2f-4966-971b-70bf271eb7a6/volumes" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.090950 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" path="/var/lib/kubelet/pods/b499bb38-4c39-4faf-9606-0ac50fe5c588/volumes" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.091760 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" path="/var/lib/kubelet/pods/d77034e9-1126-4903-8484-a99e47d2aa7f/volumes" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.092728 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" path="/var/lib/kubelet/pods/e1382c7c-c62d-46f2-ba3c-a325af801eaf/volumes" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.095883 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" path="/var/lib/kubelet/pods/fcf76dd9-1d7b-437b-9095-f5f043fce7f3/volumes" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.470759 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jlhdm"] Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471115 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471136 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471152 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471161 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471170 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471181 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471194 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471203 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471216 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471225 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471237 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471244 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471263 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471270 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471283 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471290 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471302 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471310 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471320 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471327 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471338 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471345 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="extract-content" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471355 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471362 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: E1011 03:10:01.471373 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471381 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="extract-utilities" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471501 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b499bb38-4c39-4faf-9606-0ac50fe5c588" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471518 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77034e9-1126-4903-8484-a99e47d2aa7f" containerName="marketplace-operator" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471532 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf76dd9-1d7b-437b-9095-f5f043fce7f3" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471545 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6ed923-5b2f-4966-971b-70bf271eb7a6" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.471555 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1382c7c-c62d-46f2-ba3c-a325af801eaf" containerName="registry-server" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.472613 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.478337 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.482381 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlhdm"] Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.510588 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-46rsl" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.562248 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdpw8\" (UniqueName: \"kubernetes.io/projected/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-kube-api-access-pdpw8\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.562327 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-utilities\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.562358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-catalog-content\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.663834 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-utilities\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.664589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-utilities\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.665139 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-catalog-content\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.665344 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-catalog-content\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.665565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdpw8\" (UniqueName: \"kubernetes.io/projected/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-kube-api-access-pdpw8\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.674066 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9gg"] Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.675919 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.678845 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.681945 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9gg"] Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.691847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdpw8\" (UniqueName: \"kubernetes.io/projected/0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5-kube-api-access-pdpw8\") pod \"certified-operators-jlhdm\" (UID: \"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5\") " pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.767234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w49qg\" (UniqueName: \"kubernetes.io/projected/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-kube-api-access-w49qg\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.767294 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-catalog-content\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.767359 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-utilities\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.819707 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.868930 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w49qg\" (UniqueName: \"kubernetes.io/projected/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-kube-api-access-w49qg\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.869004 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-catalog-content\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.869072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-utilities\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.869599 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-utilities\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.870124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-catalog-content\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:01 crc kubenswrapper[4754]: I1011 03:10:01.884657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w49qg\" (UniqueName: \"kubernetes.io/projected/da6b7d01-e8c6-401e-8a3c-0f9504c28b86-kube-api-access-w49qg\") pod \"redhat-marketplace-zt9gg\" (UID: \"da6b7d01-e8c6-401e-8a3c-0f9504c28b86\") " pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.001833 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlhdm"] Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.016882 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.201270 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9gg"] Oct 11 03:10:02 crc kubenswrapper[4754]: W1011 03:10:02.229309 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda6b7d01_e8c6_401e_8a3c_0f9504c28b86.slice/crio-2aec99db05889ec28a59c86c798ec24f52f1454157fd12c5ae749a769b564c04 WatchSource:0}: Error finding container 2aec99db05889ec28a59c86c798ec24f52f1454157fd12c5ae749a769b564c04: Status 404 returned error can't find the container with id 2aec99db05889ec28a59c86c798ec24f52f1454157fd12c5ae749a769b564c04 Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.500299 4754 generic.go:334] "Generic (PLEG): container finished" podID="da6b7d01-e8c6-401e-8a3c-0f9504c28b86" containerID="9cb8e85a325f0ad3dec4ee9bc18663764232e27d5fb1e9f536e47d03c95536d2" exitCode=0 Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.500398 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9gg" event={"ID":"da6b7d01-e8c6-401e-8a3c-0f9504c28b86","Type":"ContainerDied","Data":"9cb8e85a325f0ad3dec4ee9bc18663764232e27d5fb1e9f536e47d03c95536d2"} Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.500754 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9gg" event={"ID":"da6b7d01-e8c6-401e-8a3c-0f9504c28b86","Type":"ContainerStarted","Data":"2aec99db05889ec28a59c86c798ec24f52f1454157fd12c5ae749a769b564c04"} Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.504407 4754 generic.go:334] "Generic (PLEG): container finished" podID="0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5" containerID="8ff5dbeeacbaa9275bd32240acff984034d96cfe675663d201c9be727b3266b6" exitCode=0 Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.504490 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlhdm" event={"ID":"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5","Type":"ContainerDied","Data":"8ff5dbeeacbaa9275bd32240acff984034d96cfe675663d201c9be727b3266b6"} Oct 11 03:10:02 crc kubenswrapper[4754]: I1011 03:10:02.504534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlhdm" event={"ID":"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5","Type":"ContainerStarted","Data":"ca857a8b9513c10a3e967f2124c165b99effbe83d27843b2b931b468394cc0f7"} Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.512162 4754 generic.go:334] "Generic (PLEG): container finished" podID="da6b7d01-e8c6-401e-8a3c-0f9504c28b86" containerID="9c956986947e8747ae518daf2becd585827e30cedf6338fe3f574f697ed25073" exitCode=0 Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.512275 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9gg" event={"ID":"da6b7d01-e8c6-401e-8a3c-0f9504c28b86","Type":"ContainerDied","Data":"9c956986947e8747ae518daf2becd585827e30cedf6338fe3f574f697ed25073"} Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.870593 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qs976"] Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.871607 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.874203 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.888183 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs976"] Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.999026 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4rf7\" (UniqueName: \"kubernetes.io/projected/35574e71-4d0b-469d-a35e-dfbfa02664a3-kube-api-access-c4rf7\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.999162 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-catalog-content\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:03 crc kubenswrapper[4754]: I1011 03:10:03.999750 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-utilities\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.069762 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kwzzg"] Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.070752 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.074602 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.083003 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kwzzg"] Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.101323 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-utilities\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.101381 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4rf7\" (UniqueName: \"kubernetes.io/projected/35574e71-4d0b-469d-a35e-dfbfa02664a3-kube-api-access-c4rf7\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.101402 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-catalog-content\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.102016 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-catalog-content\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.102093 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35574e71-4d0b-469d-a35e-dfbfa02664a3-utilities\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.128034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4rf7\" (UniqueName: \"kubernetes.io/projected/35574e71-4d0b-469d-a35e-dfbfa02664a3-kube-api-access-c4rf7\") pod \"redhat-operators-qs976\" (UID: \"35574e71-4d0b-469d-a35e-dfbfa02664a3\") " pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.203112 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-catalog-content\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.203655 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-utilities\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.204035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmvw\" (UniqueName: \"kubernetes.io/projected/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-kube-api-access-8cmvw\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.240848 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.305913 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-utilities\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.306022 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmvw\" (UniqueName: \"kubernetes.io/projected/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-kube-api-access-8cmvw\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.306074 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-catalog-content\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.306751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-utilities\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.306783 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-catalog-content\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.329793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmvw\" (UniqueName: \"kubernetes.io/projected/54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d-kube-api-access-8cmvw\") pod \"community-operators-kwzzg\" (UID: \"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d\") " pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.394841 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.473852 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs976"] Oct 11 03:10:04 crc kubenswrapper[4754]: W1011 03:10:04.498109 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35574e71_4d0b_469d_a35e_dfbfa02664a3.slice/crio-522fd6073d4ef43da92fbb9a20a072a3b733ccc7213a1e91fecf9e599d4413bc WatchSource:0}: Error finding container 522fd6073d4ef43da92fbb9a20a072a3b733ccc7213a1e91fecf9e599d4413bc: Status 404 returned error can't find the container with id 522fd6073d4ef43da92fbb9a20a072a3b733ccc7213a1e91fecf9e599d4413bc Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.521942 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs976" event={"ID":"35574e71-4d0b-469d-a35e-dfbfa02664a3","Type":"ContainerStarted","Data":"522fd6073d4ef43da92fbb9a20a072a3b733ccc7213a1e91fecf9e599d4413bc"} Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.524131 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9gg" event={"ID":"da6b7d01-e8c6-401e-8a3c-0f9504c28b86","Type":"ContainerStarted","Data":"ea19baca384997bd07146a82dfda7dc333035c46e45b26edf41135f0fe3c1e50"} Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.537123 4754 generic.go:334] "Generic (PLEG): container finished" podID="0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5" containerID="0775e57416aa46cf0e3cd97cf95b8df5d50310fd64a16e48063b066b072f02db" exitCode=0 Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.537196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlhdm" event={"ID":"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5","Type":"ContainerDied","Data":"0775e57416aa46cf0e3cd97cf95b8df5d50310fd64a16e48063b066b072f02db"} Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.555946 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zt9gg" podStartSLOduration=2.069349639 podStartE2EDuration="3.555926791s" podCreationTimestamp="2025-10-11 03:10:01 +0000 UTC" firstStartedPulling="2025-10-11 03:10:02.501802395 +0000 UTC m=+250.060747180" lastFinishedPulling="2025-10-11 03:10:03.988379547 +0000 UTC m=+251.547324332" observedRunningTime="2025-10-11 03:10:04.553792691 +0000 UTC m=+252.112737476" watchObservedRunningTime="2025-10-11 03:10:04.555926791 +0000 UTC m=+252.114871576" Oct 11 03:10:04 crc kubenswrapper[4754]: I1011 03:10:04.615003 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kwzzg"] Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.543773 4754 generic.go:334] "Generic (PLEG): container finished" podID="35574e71-4d0b-469d-a35e-dfbfa02664a3" containerID="f7b91b6e5d165ad9bed40816a49a99d541f1220cf7bce2492d20a3a2d2ab3fa6" exitCode=0 Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.543842 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs976" event={"ID":"35574e71-4d0b-469d-a35e-dfbfa02664a3","Type":"ContainerDied","Data":"f7b91b6e5d165ad9bed40816a49a99d541f1220cf7bce2492d20a3a2d2ab3fa6"} Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.548912 4754 generic.go:334] "Generic (PLEG): container finished" podID="54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d" containerID="df52413ccb08d26e292e7fa6e3773da3c1921e6bb1bff4350f3c64e356e6ab74" exitCode=0 Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.549038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwzzg" event={"ID":"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d","Type":"ContainerDied","Data":"df52413ccb08d26e292e7fa6e3773da3c1921e6bb1bff4350f3c64e356e6ab74"} Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.549061 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwzzg" event={"ID":"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d","Type":"ContainerStarted","Data":"7908ee8d8d78158483a4a04994a52576a2ef0e0271a9589c28e63ceec50e8587"} Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.556515 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlhdm" event={"ID":"0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5","Type":"ContainerStarted","Data":"32f9080e7a44afc8113fd7c9090c6043723228b4b0c9b079add4437ee49b5348"} Oct 11 03:10:05 crc kubenswrapper[4754]: I1011 03:10:05.591684 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jlhdm" podStartSLOduration=2.092745574 podStartE2EDuration="4.591664466s" podCreationTimestamp="2025-10-11 03:10:01 +0000 UTC" firstStartedPulling="2025-10-11 03:10:02.505666583 +0000 UTC m=+250.064611378" lastFinishedPulling="2025-10-11 03:10:05.004585485 +0000 UTC m=+252.563530270" observedRunningTime="2025-10-11 03:10:05.588357527 +0000 UTC m=+253.147302312" watchObservedRunningTime="2025-10-11 03:10:05.591664466 +0000 UTC m=+253.150609251" Oct 11 03:10:07 crc kubenswrapper[4754]: I1011 03:10:07.570173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwzzg" event={"ID":"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d","Type":"ContainerStarted","Data":"0dd8333b9c05bcc7fad26aae1934917488820f2184a575a6197246aa40f006f9"} Oct 11 03:10:07 crc kubenswrapper[4754]: I1011 03:10:07.573637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs976" event={"ID":"35574e71-4d0b-469d-a35e-dfbfa02664a3","Type":"ContainerStarted","Data":"3b901042a62d23c07a943a94629bfc8b7a62a794b8a2bba5c9b023c25149fbae"} Oct 11 03:10:08 crc kubenswrapper[4754]: I1011 03:10:08.581602 4754 generic.go:334] "Generic (PLEG): container finished" podID="35574e71-4d0b-469d-a35e-dfbfa02664a3" containerID="3b901042a62d23c07a943a94629bfc8b7a62a794b8a2bba5c9b023c25149fbae" exitCode=0 Oct 11 03:10:08 crc kubenswrapper[4754]: I1011 03:10:08.581659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs976" event={"ID":"35574e71-4d0b-469d-a35e-dfbfa02664a3","Type":"ContainerDied","Data":"3b901042a62d23c07a943a94629bfc8b7a62a794b8a2bba5c9b023c25149fbae"} Oct 11 03:10:08 crc kubenswrapper[4754]: I1011 03:10:08.587922 4754 generic.go:334] "Generic (PLEG): container finished" podID="54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d" containerID="0dd8333b9c05bcc7fad26aae1934917488820f2184a575a6197246aa40f006f9" exitCode=0 Oct 11 03:10:08 crc kubenswrapper[4754]: I1011 03:10:08.587992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwzzg" event={"ID":"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d","Type":"ContainerDied","Data":"0dd8333b9c05bcc7fad26aae1934917488820f2184a575a6197246aa40f006f9"} Oct 11 03:10:09 crc kubenswrapper[4754]: I1011 03:10:09.596822 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs976" event={"ID":"35574e71-4d0b-469d-a35e-dfbfa02664a3","Type":"ContainerStarted","Data":"d807c427f02c9346236a56f81121e49581cfc28b06fd1aed2aad54c50031b363"} Oct 11 03:10:09 crc kubenswrapper[4754]: I1011 03:10:09.598909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwzzg" event={"ID":"54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d","Type":"ContainerStarted","Data":"0c6afee7157a2dfd59ec4e0b66c93490f27e36b6b51bdccb5bb065741783c05b"} Oct 11 03:10:09 crc kubenswrapper[4754]: I1011 03:10:09.615426 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qs976" podStartSLOduration=3.055851264 podStartE2EDuration="6.615411195s" podCreationTimestamp="2025-10-11 03:10:03 +0000 UTC" firstStartedPulling="2025-10-11 03:10:05.547889589 +0000 UTC m=+253.106834374" lastFinishedPulling="2025-10-11 03:10:09.10744951 +0000 UTC m=+256.666394305" observedRunningTime="2025-10-11 03:10:09.613408908 +0000 UTC m=+257.172353703" watchObservedRunningTime="2025-10-11 03:10:09.615411195 +0000 UTC m=+257.174355980" Oct 11 03:10:09 crc kubenswrapper[4754]: I1011 03:10:09.637783 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kwzzg" podStartSLOduration=2.148945293 podStartE2EDuration="5.637760374s" podCreationTimestamp="2025-10-11 03:10:04 +0000 UTC" firstStartedPulling="2025-10-11 03:10:05.550782145 +0000 UTC m=+253.109726930" lastFinishedPulling="2025-10-11 03:10:09.039597216 +0000 UTC m=+256.598542011" observedRunningTime="2025-10-11 03:10:09.636556414 +0000 UTC m=+257.195501219" watchObservedRunningTime="2025-10-11 03:10:09.637760374 +0000 UTC m=+257.196705159" Oct 11 03:10:11 crc kubenswrapper[4754]: I1011 03:10:11.820781 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:11 crc kubenswrapper[4754]: I1011 03:10:11.821236 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:11 crc kubenswrapper[4754]: I1011 03:10:11.874708 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:12 crc kubenswrapper[4754]: I1011 03:10:12.017586 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:12 crc kubenswrapper[4754]: I1011 03:10:12.017648 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:12 crc kubenswrapper[4754]: I1011 03:10:12.053698 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:12 crc kubenswrapper[4754]: I1011 03:10:12.646274 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jlhdm" Oct 11 03:10:12 crc kubenswrapper[4754]: I1011 03:10:12.654575 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zt9gg" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.241191 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.241810 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.287304 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.395977 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.396030 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.435812 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.664652 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qs976" Oct 11 03:10:14 crc kubenswrapper[4754]: I1011 03:10:14.683220 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kwzzg" Oct 11 03:11:30 crc kubenswrapper[4754]: I1011 03:11:30.735986 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:11:30 crc kubenswrapper[4754]: I1011 03:11:30.736705 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:12:00 crc kubenswrapper[4754]: I1011 03:12:00.736472 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:12:00 crc kubenswrapper[4754]: I1011 03:12:00.737266 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:12:30 crc kubenswrapper[4754]: I1011 03:12:30.736361 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:12:30 crc kubenswrapper[4754]: I1011 03:12:30.736846 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:12:30 crc kubenswrapper[4754]: I1011 03:12:30.736890 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:12:30 crc kubenswrapper[4754]: I1011 03:12:30.737319 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:12:30 crc kubenswrapper[4754]: I1011 03:12:30.737411 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd" gracePeriod=600 Oct 11 03:12:31 crc kubenswrapper[4754]: I1011 03:12:31.544612 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd" exitCode=0 Oct 11 03:12:31 crc kubenswrapper[4754]: I1011 03:12:31.544850 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd"} Oct 11 03:12:31 crc kubenswrapper[4754]: I1011 03:12:31.545221 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec"} Oct 11 03:12:31 crc kubenswrapper[4754]: I1011 03:12:31.545275 4754 scope.go:117] "RemoveContainer" containerID="c7f3b2846cd149bb37daf9bdf89eab1a56e0c65780c3cc663be82c5c8c505195" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.569041 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4v8bz"] Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.570825 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.590353 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4v8bz"] Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701598 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-trusted-ca\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701657 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-certificates\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701682 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntlll\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-kube-api-access-ntlll\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701700 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-tls\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701732 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701755 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-bound-sa-token\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.701773 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.721726 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-certificates\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntlll\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-kube-api-access-ntlll\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-tls\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803566 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-bound-sa-token\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803601 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803654 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.803673 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-trusted-ca\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.804375 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-certificates\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.804645 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.804729 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-trusted-ca\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.810077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-registry-tls\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.810128 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.819639 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-bound-sa-token\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.823036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntlll\" (UniqueName: \"kubernetes.io/projected/d9f6c667-a50e-4e3e-acab-7f0a424a43a3-kube-api-access-ntlll\") pod \"image-registry-66df7c8f76-4v8bz\" (UID: \"d9f6c667-a50e-4e3e-acab-7f0a424a43a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:13:59 crc kubenswrapper[4754]: I1011 03:13:59.888800 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:14:00 crc kubenswrapper[4754]: I1011 03:14:00.084999 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4v8bz"] Oct 11 03:14:00 crc kubenswrapper[4754]: I1011 03:14:00.129007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" event={"ID":"d9f6c667-a50e-4e3e-acab-7f0a424a43a3","Type":"ContainerStarted","Data":"1944e986c076bc2a8366afaa6f78a9a271980b741ec33be97fbc405dc70cba79"} Oct 11 03:14:01 crc kubenswrapper[4754]: I1011 03:14:01.140891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" event={"ID":"d9f6c667-a50e-4e3e-acab-7f0a424a43a3","Type":"ContainerStarted","Data":"a207374babd19e2fd71af2fcb57e53197609ff85f3cba08a2a529ce84ba11791"} Oct 11 03:14:01 crc kubenswrapper[4754]: I1011 03:14:01.141170 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:14:01 crc kubenswrapper[4754]: I1011 03:14:01.165122 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" podStartSLOduration=2.165095601 podStartE2EDuration="2.165095601s" podCreationTimestamp="2025-10-11 03:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:14:01.161326944 +0000 UTC m=+488.720271789" watchObservedRunningTime="2025-10-11 03:14:01.165095601 +0000 UTC m=+488.724040426" Oct 11 03:14:19 crc kubenswrapper[4754]: I1011 03:14:19.893596 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4v8bz" Oct 11 03:14:19 crc kubenswrapper[4754]: I1011 03:14:19.961735 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.006178 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" podUID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" containerName="registry" containerID="cri-o://1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd" gracePeriod=30 Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.401318 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420127 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420203 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420245 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420336 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66pd7\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420390 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.420439 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.424945 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.425875 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.433645 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7" (OuterVolumeSpecName: "kube-api-access-66pd7") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "kube-api-access-66pd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.433731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.434149 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.439665 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.500761 4754 generic.go:334] "Generic (PLEG): container finished" podID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" containerID="1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd" exitCode=0 Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.500821 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" event={"ID":"4eefc33f-2e4f-4bd2-9866-2fc103a44135","Type":"ContainerDied","Data":"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd"} Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.500862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" event={"ID":"4eefc33f-2e4f-4bd2-9866-2fc103a44135","Type":"ContainerDied","Data":"38e360f4862989cf32df1b071d4d48b257987bb45071ef0e9387c6ab4fe0455a"} Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.500885 4754 scope.go:117] "RemoveContainer" containerID="1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.501070 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx5qd" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521089 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521138 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted\") pod \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\" (UID: \"4eefc33f-2e4f-4bd2-9866-2fc103a44135\") " Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521352 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4eefc33f-2e4f-4bd2-9866-2fc103a44135-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521371 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521381 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eefc33f-2e4f-4bd2-9866-2fc103a44135-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521392 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66pd7\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-kube-api-access-66pd7\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.521404 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.522217 4754 scope.go:117] "RemoveContainer" containerID="1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd" Oct 11 03:14:45 crc kubenswrapper[4754]: E1011 03:14:45.523500 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd\": container with ID starting with 1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd not found: ID does not exist" containerID="1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.523537 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd"} err="failed to get container status \"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd\": rpc error: code = NotFound desc = could not find container \"1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd\": container with ID starting with 1744c795d556899bf89de80082e36f7318ea0345f73dd85c211c8aea66f011dd not found: ID does not exist" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.525016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.547882 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4eefc33f-2e4f-4bd2-9866-2fc103a44135" (UID: "4eefc33f-2e4f-4bd2-9866-2fc103a44135"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.622507 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4eefc33f-2e4f-4bd2-9866-2fc103a44135-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.622546 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4eefc33f-2e4f-4bd2-9866-2fc103a44135-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.829502 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:14:45 crc kubenswrapper[4754]: I1011 03:14:45.842130 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx5qd"] Oct 11 03:14:47 crc kubenswrapper[4754]: I1011 03:14:47.097489 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" path="/var/lib/kubelet/pods/4eefc33f-2e4f-4bd2-9866-2fc103a44135/volumes" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.164195 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq"] Oct 11 03:15:00 crc kubenswrapper[4754]: E1011 03:15:00.164995 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" containerName="registry" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.165009 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" containerName="registry" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.165108 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eefc33f-2e4f-4bd2-9866-2fc103a44135" containerName="registry" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.165466 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.168783 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.168838 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.170834 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq"] Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.328578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.328634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.328661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrjq6\" (UniqueName: \"kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.430089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.430157 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.430184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrjq6\" (UniqueName: \"kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.431272 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.436030 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.444278 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrjq6\" (UniqueName: \"kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6\") pod \"collect-profiles-29335875-qxtrq\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.488180 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.657132 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq"] Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.736778 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:15:00 crc kubenswrapper[4754]: I1011 03:15:00.736857 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:15:01 crc kubenswrapper[4754]: I1011 03:15:01.617228 4754 generic.go:334] "Generic (PLEG): container finished" podID="e9bc4270-b249-44eb-9931-b026cf1721f0" containerID="19ba11560fc0ad36ba3b52afbb484b0afc1a78b78c599b27fc84fd47db75a773" exitCode=0 Oct 11 03:15:01 crc kubenswrapper[4754]: I1011 03:15:01.617476 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" event={"ID":"e9bc4270-b249-44eb-9931-b026cf1721f0","Type":"ContainerDied","Data":"19ba11560fc0ad36ba3b52afbb484b0afc1a78b78c599b27fc84fd47db75a773"} Oct 11 03:15:01 crc kubenswrapper[4754]: I1011 03:15:01.617677 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" event={"ID":"e9bc4270-b249-44eb-9931-b026cf1721f0","Type":"ContainerStarted","Data":"2a238e71d3aaa2307ba44130825c474e55dcc41d01a636eeb39bc1d146fafba8"} Oct 11 03:15:02 crc kubenswrapper[4754]: I1011 03:15:02.962128 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.062315 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume\") pod \"e9bc4270-b249-44eb-9931-b026cf1721f0\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.062495 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrjq6\" (UniqueName: \"kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6\") pod \"e9bc4270-b249-44eb-9931-b026cf1721f0\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.062568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume\") pod \"e9bc4270-b249-44eb-9931-b026cf1721f0\" (UID: \"e9bc4270-b249-44eb-9931-b026cf1721f0\") " Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.063676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9bc4270-b249-44eb-9931-b026cf1721f0" (UID: "e9bc4270-b249-44eb-9931-b026cf1721f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.069332 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6" (OuterVolumeSpecName: "kube-api-access-nrjq6") pod "e9bc4270-b249-44eb-9931-b026cf1721f0" (UID: "e9bc4270-b249-44eb-9931-b026cf1721f0"). InnerVolumeSpecName "kube-api-access-nrjq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.070278 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9bc4270-b249-44eb-9931-b026cf1721f0" (UID: "e9bc4270-b249-44eb-9931-b026cf1721f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.164318 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrjq6\" (UniqueName: \"kubernetes.io/projected/e9bc4270-b249-44eb-9931-b026cf1721f0-kube-api-access-nrjq6\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.164634 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9bc4270-b249-44eb-9931-b026cf1721f0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.164727 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9bc4270-b249-44eb-9931-b026cf1721f0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.631089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" event={"ID":"e9bc4270-b249-44eb-9931-b026cf1721f0","Type":"ContainerDied","Data":"2a238e71d3aaa2307ba44130825c474e55dcc41d01a636eeb39bc1d146fafba8"} Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.631135 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a238e71d3aaa2307ba44130825c474e55dcc41d01a636eeb39bc1d146fafba8" Oct 11 03:15:03 crc kubenswrapper[4754]: I1011 03:15:03.631192 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.344462 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p6m9q"] Oct 11 03:15:04 crc kubenswrapper[4754]: E1011 03:15:04.344665 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bc4270-b249-44eb-9931-b026cf1721f0" containerName="collect-profiles" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.344677 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bc4270-b249-44eb-9931-b026cf1721f0" containerName="collect-profiles" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.344776 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bc4270-b249-44eb-9931-b026cf1721f0" containerName="collect-profiles" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.345235 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.364387 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.364504 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.366419 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-fjczj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.367776 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p6m9q"] Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.370001 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6j2fj"] Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.370600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6j2fj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.372114 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-m5xhd" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.391461 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6j2fj"] Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.395048 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g6lpk"] Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.395814 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.398003 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tq82r" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.411464 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g6lpk"] Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.481036 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrlsx\" (UniqueName: \"kubernetes.io/projected/04bdf1ce-f0a8-4b76-870b-b513547b391a-kube-api-access-zrlsx\") pod \"cert-manager-5b446d88c5-6j2fj\" (UID: \"04bdf1ce-f0a8-4b76-870b-b513547b391a\") " pod="cert-manager/cert-manager-5b446d88c5-6j2fj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.481390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfgfz\" (UniqueName: \"kubernetes.io/projected/14fb2f1e-9c40-462d-bd5e-db8bdcd0db23-kube-api-access-kfgfz\") pod \"cert-manager-cainjector-7f985d654d-p6m9q\" (UID: \"14fb2f1e-9c40-462d-bd5e-db8bdcd0db23\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.481540 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9zrv\" (UniqueName: \"kubernetes.io/projected/f4adad4f-1d97-4526-9d87-01552818cb64-kube-api-access-l9zrv\") pod \"cert-manager-webhook-5655c58dd6-g6lpk\" (UID: \"f4adad4f-1d97-4526-9d87-01552818cb64\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.582775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrlsx\" (UniqueName: \"kubernetes.io/projected/04bdf1ce-f0a8-4b76-870b-b513547b391a-kube-api-access-zrlsx\") pod \"cert-manager-5b446d88c5-6j2fj\" (UID: \"04bdf1ce-f0a8-4b76-870b-b513547b391a\") " pod="cert-manager/cert-manager-5b446d88c5-6j2fj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.582822 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfgfz\" (UniqueName: \"kubernetes.io/projected/14fb2f1e-9c40-462d-bd5e-db8bdcd0db23-kube-api-access-kfgfz\") pod \"cert-manager-cainjector-7f985d654d-p6m9q\" (UID: \"14fb2f1e-9c40-462d-bd5e-db8bdcd0db23\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.582864 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9zrv\" (UniqueName: \"kubernetes.io/projected/f4adad4f-1d97-4526-9d87-01552818cb64-kube-api-access-l9zrv\") pod \"cert-manager-webhook-5655c58dd6-g6lpk\" (UID: \"f4adad4f-1d97-4526-9d87-01552818cb64\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.602410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfgfz\" (UniqueName: \"kubernetes.io/projected/14fb2f1e-9c40-462d-bd5e-db8bdcd0db23-kube-api-access-kfgfz\") pod \"cert-manager-cainjector-7f985d654d-p6m9q\" (UID: \"14fb2f1e-9c40-462d-bd5e-db8bdcd0db23\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.607259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrlsx\" (UniqueName: \"kubernetes.io/projected/04bdf1ce-f0a8-4b76-870b-b513547b391a-kube-api-access-zrlsx\") pod \"cert-manager-5b446d88c5-6j2fj\" (UID: \"04bdf1ce-f0a8-4b76-870b-b513547b391a\") " pod="cert-manager/cert-manager-5b446d88c5-6j2fj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.608099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9zrv\" (UniqueName: \"kubernetes.io/projected/f4adad4f-1d97-4526-9d87-01552818cb64-kube-api-access-l9zrv\") pod \"cert-manager-webhook-5655c58dd6-g6lpk\" (UID: \"f4adad4f-1d97-4526-9d87-01552818cb64\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.661236 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.687107 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6j2fj" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.714516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.972617 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p6m9q"] Oct 11 03:15:04 crc kubenswrapper[4754]: W1011 03:15:04.979006 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14fb2f1e_9c40_462d_bd5e_db8bdcd0db23.slice/crio-16cf853a37b79ff1526fc0525cae47e58d93033ac2598594e8e76f2f3526cff7 WatchSource:0}: Error finding container 16cf853a37b79ff1526fc0525cae47e58d93033ac2598594e8e76f2f3526cff7: Status 404 returned error can't find the container with id 16cf853a37b79ff1526fc0525cae47e58d93033ac2598594e8e76f2f3526cff7 Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.981777 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:15:04 crc kubenswrapper[4754]: I1011 03:15:04.982669 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g6lpk"] Oct 11 03:15:05 crc kubenswrapper[4754]: I1011 03:15:05.144408 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6j2fj"] Oct 11 03:15:05 crc kubenswrapper[4754]: I1011 03:15:05.643622 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" event={"ID":"14fb2f1e-9c40-462d-bd5e-db8bdcd0db23","Type":"ContainerStarted","Data":"16cf853a37b79ff1526fc0525cae47e58d93033ac2598594e8e76f2f3526cff7"} Oct 11 03:15:05 crc kubenswrapper[4754]: I1011 03:15:05.644654 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" event={"ID":"f4adad4f-1d97-4526-9d87-01552818cb64","Type":"ContainerStarted","Data":"1e9360b8daf33a49a69d260adcad3a3018cc0ba49d81310e1e13a6697387abe7"} Oct 11 03:15:05 crc kubenswrapper[4754]: I1011 03:15:05.645821 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6j2fj" event={"ID":"04bdf1ce-f0a8-4b76-870b-b513547b391a","Type":"ContainerStarted","Data":"86d3ac6de0a7c8f73f7bf4d8922e09620d7376fb51b449bb2cc459565304ed07"} Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.663015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" event={"ID":"f4adad4f-1d97-4526-9d87-01552818cb64","Type":"ContainerStarted","Data":"8cdef152679a36697093d9cf32a9cbb27377294185c1aea1953626fd340ea400"} Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.663440 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.664923 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6j2fj" event={"ID":"04bdf1ce-f0a8-4b76-870b-b513547b391a","Type":"ContainerStarted","Data":"3d84041c0ccdda4a4033847cca7e6ac99b9d12428937ab5146641522eed70391"} Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.666318 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" event={"ID":"14fb2f1e-9c40-462d-bd5e-db8bdcd0db23","Type":"ContainerStarted","Data":"24014a2ab0d441911a2f3c812a2707a092029a1d13ef4253993ede2439467be5"} Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.681112 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" podStartSLOduration=1.723565354 podStartE2EDuration="4.681081567s" podCreationTimestamp="2025-10-11 03:15:04 +0000 UTC" firstStartedPulling="2025-10-11 03:15:04.988723396 +0000 UTC m=+552.547668181" lastFinishedPulling="2025-10-11 03:15:07.946239599 +0000 UTC m=+555.505184394" observedRunningTime="2025-10-11 03:15:08.679019229 +0000 UTC m=+556.237964014" watchObservedRunningTime="2025-10-11 03:15:08.681081567 +0000 UTC m=+556.240026392" Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.695310 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-6j2fj" podStartSLOduration=1.8196249660000001 podStartE2EDuration="4.695277806s" podCreationTimestamp="2025-10-11 03:15:04 +0000 UTC" firstStartedPulling="2025-10-11 03:15:05.15522351 +0000 UTC m=+552.714168295" lastFinishedPulling="2025-10-11 03:15:08.03087635 +0000 UTC m=+555.589821135" observedRunningTime="2025-10-11 03:15:08.692706424 +0000 UTC m=+556.251651229" watchObservedRunningTime="2025-10-11 03:15:08.695277806 +0000 UTC m=+556.254222601" Oct 11 03:15:08 crc kubenswrapper[4754]: I1011 03:15:08.714434 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-p6m9q" podStartSLOduration=1.7506060940000001 podStartE2EDuration="4.714387133s" podCreationTimestamp="2025-10-11 03:15:04 +0000 UTC" firstStartedPulling="2025-10-11 03:15:04.981526674 +0000 UTC m=+552.540471459" lastFinishedPulling="2025-10-11 03:15:07.945307713 +0000 UTC m=+555.504252498" observedRunningTime="2025-10-11 03:15:08.704479025 +0000 UTC m=+556.263423840" watchObservedRunningTime="2025-10-11 03:15:08.714387133 +0000 UTC m=+556.273331918" Oct 11 03:15:14 crc kubenswrapper[4754]: I1011 03:15:14.719279 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-g6lpk" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.081509 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-42sbf"] Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.081911 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-controller" containerID="cri-o://cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082013 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082067 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-node" containerID="cri-o://bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082004 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="nbdb" containerID="cri-o://3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082114 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-acl-logging" containerID="cri-o://48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082445 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="northd" containerID="cri-o://47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.082511 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="sbdb" containerID="cri-o://9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.119767 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" containerID="cri-o://4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" gracePeriod=30 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.421848 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/3.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.424754 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovn-acl-logging/0.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.425206 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovn-controller/0.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.425598 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482436 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtxbg"] Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482684 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-node" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482704 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-node" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482722 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="nbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482731 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="nbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482743 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482751 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482760 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482768 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482778 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482791 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482805 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482817 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482831 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="northd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482839 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="northd" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482848 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="sbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482856 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="sbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482868 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482879 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482898 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kubecfg-setup" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482909 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kubecfg-setup" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482923 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-acl-logging" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482931 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-acl-logging" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.482944 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.482952 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483089 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483104 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-node" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483121 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483133 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="northd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483148 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovn-acl-logging" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483163 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483173 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483185 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="nbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483194 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483207 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483217 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="sbdb" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.483377 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483391 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.483528 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerName="ovnkube-controller" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.485776 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533192 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533251 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533276 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533327 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533363 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533384 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533407 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533424 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533453 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533492 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533540 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533567 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533588 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4t5f\" (UniqueName: \"kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533607 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533631 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533659 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533691 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.533715 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib\") pod \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\" (UID: \"0dea8260-bc1e-4ad6-96d3-f9d202b125c7\") " Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534254 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534330 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534333 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534365 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534409 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534435 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534459 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534797 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log" (OuterVolumeSpecName: "node-log") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534823 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.534873 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535183 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535230 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535265 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash" (OuterVolumeSpecName: "host-slash") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535330 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket" (OuterVolumeSpecName: "log-socket") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.535356 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.543495 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.543848 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f" (OuterVolumeSpecName: "kube-api-access-v4t5f") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "kube-api-access-v4t5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.550511 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0dea8260-bc1e-4ad6-96d3-f9d202b125c7" (UID: "0dea8260-bc1e-4ad6-96d3-f9d202b125c7"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.634981 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-kubelet\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635023 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovn-node-metrics-cert\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-ovn\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635117 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-netns\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635137 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-node-log\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635152 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-bin\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-env-overrides\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635219 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-systemd-units\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635237 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-netd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635254 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-script-lib\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnhzv\" (UniqueName: \"kubernetes.io/projected/a6d406ca-40b8-4991-b6a1-a1ba5199c303-kube-api-access-hnhzv\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635400 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-etc-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635424 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-log-socket\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635508 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635548 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-config\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635562 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-slash\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635583 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-systemd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-var-lib-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635767 4754 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635796 4754 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635806 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635816 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635825 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635833 4754 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635842 4754 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635851 4754 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635860 4754 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635869 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4t5f\" (UniqueName: \"kubernetes.io/projected/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-kube-api-access-v4t5f\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635877 4754 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635885 4754 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-slash\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635894 4754 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-log-socket\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635901 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635909 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635917 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635927 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635934 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635942 4754 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-node-log\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.635950 4754 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dea8260-bc1e-4ad6-96d3-f9d202b125c7-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.710346 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/2.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.710796 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/1.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.710834 4754 generic.go:334] "Generic (PLEG): container finished" podID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" containerID="c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf" exitCode=2 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.710903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerDied","Data":"c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.710955 4754 scope.go:117] "RemoveContainer" containerID="5ff219ddf626f99c6b4214ff6d8224958747ba4a40b0afeb3a3f7adc36a22cd2" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.711432 4754 scope.go:117] "RemoveContainer" containerID="c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.711630 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2nlvv_openshift-multus(3cd26d5b-a1b6-41fa-b285-d650e389d8d4)\"" pod="openshift-multus/multus-2nlvv" podUID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.713589 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovnkube-controller/3.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.715845 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovn-acl-logging/0.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716352 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-42sbf_0dea8260-bc1e-4ad6-96d3-f9d202b125c7/ovn-controller/0.log" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716730 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716749 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716756 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716765 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716757 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716807 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716822 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716855 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716773 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716975 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" exitCode=0 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.716995 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" exitCode=143 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717005 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" exitCode=143 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717023 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717041 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717055 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717063 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717070 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717077 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717085 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717092 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717100 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717108 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717116 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717126 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717139 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717147 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717154 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717161 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717168 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717176 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717183 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717190 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717197 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717204 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717229 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717238 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717245 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717253 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717259 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717266 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717273 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717280 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717286 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717292 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717301 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-42sbf" event={"ID":"0dea8260-bc1e-4ad6-96d3-f9d202b125c7","Type":"ContainerDied","Data":"3fe3d31161b7a0f6a9dfb1bf95acfe17ad3d904f7941e3cb2aa64a7294ce0c80"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717312 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717319 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717327 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717334 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717341 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717348 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717354 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717362 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717368 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.717376 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736668 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovn-node-metrics-cert\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736723 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-ovn\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-netns\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-node-log\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736799 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-bin\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736819 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-env-overrides\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736846 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-systemd-units\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-netd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-script-lib\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnhzv\" (UniqueName: \"kubernetes.io/projected/a6d406ca-40b8-4991-b6a1-a1ba5199c303-kube-api-access-hnhzv\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.736987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-log-socket\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737009 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-etc-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-config\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-slash\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737115 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-systemd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-var-lib-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737175 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-kubelet\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737278 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-kubelet\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737574 4754 scope.go:117] "RemoveContainer" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-bin\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737911 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-ovn\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.737942 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-netns\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-systemd-units\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-node-log\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738013 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-log-socket\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738109 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-systemd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-slash\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-etc-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738235 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-run-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-var-lib-openvswitch\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738316 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-cni-netd\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738339 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738319 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a6d406ca-40b8-4991-b6a1-a1ba5199c303-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738533 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-env-overrides\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-config\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.738555 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovnkube-script-lib\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.741734 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6d406ca-40b8-4991-b6a1-a1ba5199c303-ovn-node-metrics-cert\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.749149 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-42sbf"] Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.756194 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-42sbf"] Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.757729 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.758029 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnhzv\" (UniqueName: \"kubernetes.io/projected/a6d406ca-40b8-4991-b6a1-a1ba5199c303-kube-api-access-hnhzv\") pod \"ovnkube-node-rtxbg\" (UID: \"a6d406ca-40b8-4991-b6a1-a1ba5199c303\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.772609 4754 scope.go:117] "RemoveContainer" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.785822 4754 scope.go:117] "RemoveContainer" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.798509 4754 scope.go:117] "RemoveContainer" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.798932 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.816875 4754 scope.go:117] "RemoveContainer" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: W1011 03:15:15.829446 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6d406ca_40b8_4991_b6a1_a1ba5199c303.slice/crio-0b4e5dc8b7b66c1215b4875d5f7758efac6c1195a7de67a84c7f28457f96bb26 WatchSource:0}: Error finding container 0b4e5dc8b7b66c1215b4875d5f7758efac6c1195a7de67a84c7f28457f96bb26: Status 404 returned error can't find the container with id 0b4e5dc8b7b66c1215b4875d5f7758efac6c1195a7de67a84c7f28457f96bb26 Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.836783 4754 scope.go:117] "RemoveContainer" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.856741 4754 scope.go:117] "RemoveContainer" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.877573 4754 scope.go:117] "RemoveContainer" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.896430 4754 scope.go:117] "RemoveContainer" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.913775 4754 scope.go:117] "RemoveContainer" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.914840 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": container with ID starting with 4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad not found: ID does not exist" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.914873 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} err="failed to get container status \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": rpc error: code = NotFound desc = could not find container \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": container with ID starting with 4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.914896 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.915417 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": container with ID starting with 90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce not found: ID does not exist" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.915446 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} err="failed to get container status \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": rpc error: code = NotFound desc = could not find container \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": container with ID starting with 90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.915462 4754 scope.go:117] "RemoveContainer" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.915670 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": container with ID starting with 9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227 not found: ID does not exist" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.915692 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} err="failed to get container status \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": rpc error: code = NotFound desc = could not find container \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": container with ID starting with 9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.915703 4754 scope.go:117] "RemoveContainer" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.916348 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": container with ID starting with 3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098 not found: ID does not exist" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916371 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} err="failed to get container status \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": rpc error: code = NotFound desc = could not find container \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": container with ID starting with 3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916385 4754 scope.go:117] "RemoveContainer" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.916568 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": container with ID starting with 47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6 not found: ID does not exist" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916588 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} err="failed to get container status \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": rpc error: code = NotFound desc = could not find container \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": container with ID starting with 47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916601 4754 scope.go:117] "RemoveContainer" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.916836 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": container with ID starting with c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd not found: ID does not exist" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916856 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} err="failed to get container status \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": rpc error: code = NotFound desc = could not find container \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": container with ID starting with c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.916867 4754 scope.go:117] "RemoveContainer" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.917620 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": container with ID starting with bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf not found: ID does not exist" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.917642 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} err="failed to get container status \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": rpc error: code = NotFound desc = could not find container \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": container with ID starting with bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.917655 4754 scope.go:117] "RemoveContainer" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.917859 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": container with ID starting with 48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558 not found: ID does not exist" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.917879 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} err="failed to get container status \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": rpc error: code = NotFound desc = could not find container \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": container with ID starting with 48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.917892 4754 scope.go:117] "RemoveContainer" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.918162 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": container with ID starting with cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7 not found: ID does not exist" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918179 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} err="failed to get container status \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": rpc error: code = NotFound desc = could not find container \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": container with ID starting with cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918192 4754 scope.go:117] "RemoveContainer" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: E1011 03:15:15.918392 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": container with ID starting with e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5 not found: ID does not exist" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918412 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} err="failed to get container status \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": rpc error: code = NotFound desc = could not find container \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": container with ID starting with e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918426 4754 scope.go:117] "RemoveContainer" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918607 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} err="failed to get container status \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": rpc error: code = NotFound desc = could not find container \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": container with ID starting with 4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918622 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918880 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} err="failed to get container status \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": rpc error: code = NotFound desc = could not find container \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": container with ID starting with 90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.918895 4754 scope.go:117] "RemoveContainer" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919086 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} err="failed to get container status \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": rpc error: code = NotFound desc = could not find container \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": container with ID starting with 9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919102 4754 scope.go:117] "RemoveContainer" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919250 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} err="failed to get container status \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": rpc error: code = NotFound desc = could not find container \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": container with ID starting with 3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919265 4754 scope.go:117] "RemoveContainer" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919524 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} err="failed to get container status \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": rpc error: code = NotFound desc = could not find container \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": container with ID starting with 47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919540 4754 scope.go:117] "RemoveContainer" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919723 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} err="failed to get container status \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": rpc error: code = NotFound desc = could not find container \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": container with ID starting with c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.919740 4754 scope.go:117] "RemoveContainer" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920002 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} err="failed to get container status \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": rpc error: code = NotFound desc = could not find container \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": container with ID starting with bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920016 4754 scope.go:117] "RemoveContainer" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920222 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} err="failed to get container status \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": rpc error: code = NotFound desc = could not find container \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": container with ID starting with 48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920240 4754 scope.go:117] "RemoveContainer" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920389 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} err="failed to get container status \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": rpc error: code = NotFound desc = could not find container \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": container with ID starting with cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920428 4754 scope.go:117] "RemoveContainer" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920592 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} err="failed to get container status \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": rpc error: code = NotFound desc = could not find container \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": container with ID starting with e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920608 4754 scope.go:117] "RemoveContainer" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920799 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} err="failed to get container status \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": rpc error: code = NotFound desc = could not find container \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": container with ID starting with 4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920819 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.920989 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} err="failed to get container status \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": rpc error: code = NotFound desc = could not find container \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": container with ID starting with 90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.921005 4754 scope.go:117] "RemoveContainer" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.921685 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} err="failed to get container status \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": rpc error: code = NotFound desc = could not find container \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": container with ID starting with 9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.921703 4754 scope.go:117] "RemoveContainer" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922020 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} err="failed to get container status \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": rpc error: code = NotFound desc = could not find container \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": container with ID starting with 3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922076 4754 scope.go:117] "RemoveContainer" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922314 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} err="failed to get container status \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": rpc error: code = NotFound desc = could not find container \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": container with ID starting with 47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922337 4754 scope.go:117] "RemoveContainer" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922563 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} err="failed to get container status \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": rpc error: code = NotFound desc = could not find container \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": container with ID starting with c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922584 4754 scope.go:117] "RemoveContainer" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922763 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} err="failed to get container status \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": rpc error: code = NotFound desc = could not find container \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": container with ID starting with bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.922795 4754 scope.go:117] "RemoveContainer" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.923695 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} err="failed to get container status \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": rpc error: code = NotFound desc = could not find container \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": container with ID starting with 48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.923728 4754 scope.go:117] "RemoveContainer" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.923994 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} err="failed to get container status \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": rpc error: code = NotFound desc = could not find container \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": container with ID starting with cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.924014 4754 scope.go:117] "RemoveContainer" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.924241 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} err="failed to get container status \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": rpc error: code = NotFound desc = could not find container \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": container with ID starting with e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.924264 4754 scope.go:117] "RemoveContainer" containerID="4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.924788 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad"} err="failed to get container status \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": rpc error: code = NotFound desc = could not find container \"4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad\": container with ID starting with 4118d5fb9c68c64c4ec73b27f212122916fbdd20e34235332a750bd4fd4aa5ad not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.924814 4754 scope.go:117] "RemoveContainer" containerID="90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.925658 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce"} err="failed to get container status \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": rpc error: code = NotFound desc = could not find container \"90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce\": container with ID starting with 90d120cd4e0bceedf264fcba7f8e61912c7676d4d1c11e0c28ccea35ee56e1ce not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.925679 4754 scope.go:117] "RemoveContainer" containerID="9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.925888 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227"} err="failed to get container status \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": rpc error: code = NotFound desc = could not find container \"9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227\": container with ID starting with 9f970cd0b5dfcf59370ced8dc3a792abe0c92d169ae52e5e8faf55a2db44b227 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.925904 4754 scope.go:117] "RemoveContainer" containerID="3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926154 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098"} err="failed to get container status \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": rpc error: code = NotFound desc = could not find container \"3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098\": container with ID starting with 3f3f5630f75bc8f3f8233dce56501dff8cf9bbae90776ce2d2f6578262257098 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926207 4754 scope.go:117] "RemoveContainer" containerID="47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926507 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6"} err="failed to get container status \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": rpc error: code = NotFound desc = could not find container \"47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6\": container with ID starting with 47f7133c66a3fb25ec567995fa61892679fdf836034fda30410f744ea621bfa6 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926525 4754 scope.go:117] "RemoveContainer" containerID="c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926687 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd"} err="failed to get container status \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": rpc error: code = NotFound desc = could not find container \"c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd\": container with ID starting with c171956811fbbcf476ef3b9bd27cf7d1b77c2b02279bd763e4d963abffd2abcd not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926712 4754 scope.go:117] "RemoveContainer" containerID="bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926887 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf"} err="failed to get container status \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": rpc error: code = NotFound desc = could not find container \"bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf\": container with ID starting with bbc46b8c7b0e3a50a9f208758fda869a7a30159760ed683f25f10f09faa3a5bf not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.926904 4754 scope.go:117] "RemoveContainer" containerID="48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.927103 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558"} err="failed to get container status \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": rpc error: code = NotFound desc = could not find container \"48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558\": container with ID starting with 48375af338e2e3391fcac9ca3d586b52dcb819a59c4f380c8bd812dc6d067558 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.927120 4754 scope.go:117] "RemoveContainer" containerID="cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.927267 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7"} err="failed to get container status \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": rpc error: code = NotFound desc = could not find container \"cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7\": container with ID starting with cb2125dd806b46ef1fee6b0dc33396d41cec899e9feac521702598dca5df8ab7 not found: ID does not exist" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.927283 4754 scope.go:117] "RemoveContainer" containerID="e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5" Oct 11 03:15:15 crc kubenswrapper[4754]: I1011 03:15:15.927441 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5"} err="failed to get container status \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": rpc error: code = NotFound desc = could not find container \"e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5\": container with ID starting with e2f296d5dea1b9c89b2c291a9fb6f703653511b50fb54fe00a71851e20c807c5 not found: ID does not exist" Oct 11 03:15:16 crc kubenswrapper[4754]: I1011 03:15:16.724119 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/2.log" Oct 11 03:15:16 crc kubenswrapper[4754]: I1011 03:15:16.726165 4754 generic.go:334] "Generic (PLEG): container finished" podID="a6d406ca-40b8-4991-b6a1-a1ba5199c303" containerID="036b046445bb4bbaeb535bbedd415aaaa002e45945a290849d422f04a67aa70d" exitCode=0 Oct 11 03:15:16 crc kubenswrapper[4754]: I1011 03:15:16.726208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerDied","Data":"036b046445bb4bbaeb535bbedd415aaaa002e45945a290849d422f04a67aa70d"} Oct 11 03:15:16 crc kubenswrapper[4754]: I1011 03:15:16.726235 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"0b4e5dc8b7b66c1215b4875d5f7758efac6c1195a7de67a84c7f28457f96bb26"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.094624 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dea8260-bc1e-4ad6-96d3-f9d202b125c7" path="/var/lib/kubelet/pods/0dea8260-bc1e-4ad6-96d3-f9d202b125c7/volumes" Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"e5c954a612060a09bcc8c2a847987a1c775b3a2fc32de7d2e5f7f910735aa0c7"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737449 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"1aa4de6f31f3ed9c3ab9e71c957ce5145f8bd982a4845ec89e444646b998c6b0"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"f284fbd45f6d1c40f8b083772c51bc568181b7aa64f71e66e3b01b9c0afca96e"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"acfb44e306ef35f6a96e9d600f8b7d2875057883c6bbbd5e23c49721b213f310"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737521 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"4d77d3e2dd711024c49199a574a4ec189d1ccfc60b94f10254fd69c9422c7204"} Oct 11 03:15:17 crc kubenswrapper[4754]: I1011 03:15:17.737538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"f79a39dff1b87e71dfd99e9d7100ebf487c30a1ab0ae9456b7ff900bb4cbbb35"} Oct 11 03:15:20 crc kubenswrapper[4754]: I1011 03:15:20.764141 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"a1e4aa59d7b25acdddbe7d1927d830b988dd0e2d1ab03d077474cd8eaca3f6bf"} Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.780087 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" event={"ID":"a6d406ca-40b8-4991-b6a1-a1ba5199c303","Type":"ContainerStarted","Data":"28f683aaffb7302ee5cdc0c914724844dbcbd730d0cc3a1ca000c7ed350c11c0"} Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.781039 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.781058 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.781072 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.808371 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.825746 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:22 crc kubenswrapper[4754]: I1011 03:15:22.828990 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" podStartSLOduration=7.828945275 podStartE2EDuration="7.828945275s" podCreationTimestamp="2025-10-11 03:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:15:22.821174156 +0000 UTC m=+570.380118961" watchObservedRunningTime="2025-10-11 03:15:22.828945275 +0000 UTC m=+570.387890060" Oct 11 03:15:30 crc kubenswrapper[4754]: I1011 03:15:30.736365 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:15:30 crc kubenswrapper[4754]: I1011 03:15:30.737488 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:15:31 crc kubenswrapper[4754]: I1011 03:15:31.083628 4754 scope.go:117] "RemoveContainer" containerID="c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf" Oct 11 03:15:31 crc kubenswrapper[4754]: E1011 03:15:31.083821 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2nlvv_openshift-multus(3cd26d5b-a1b6-41fa-b285-d650e389d8d4)\"" pod="openshift-multus/multus-2nlvv" podUID="3cd26d5b-a1b6-41fa-b285-d650e389d8d4" Oct 11 03:15:42 crc kubenswrapper[4754]: I1011 03:15:42.083565 4754 scope.go:117] "RemoveContainer" containerID="c7402700a1a1282d874956b6eeaf344f6edaaa742d8661368b1469e7f6adb5bf" Oct 11 03:15:42 crc kubenswrapper[4754]: I1011 03:15:42.920117 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2nlvv_3cd26d5b-a1b6-41fa-b285-d650e389d8d4/kube-multus/2.log" Oct 11 03:15:42 crc kubenswrapper[4754]: I1011 03:15:42.921054 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2nlvv" event={"ID":"3cd26d5b-a1b6-41fa-b285-d650e389d8d4","Type":"ContainerStarted","Data":"4eff61b43ec0bbde526fbc353050ea321751e7721061fe60b1cd721d9ea0cb35"} Oct 11 03:15:45 crc kubenswrapper[4754]: I1011 03:15:45.863920 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtxbg" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.492516 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs"] Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.494094 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.504082 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.525193 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs"] Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.589474 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc95g\" (UniqueName: \"kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.589554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.589630 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.691103 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.691174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc95g\" (UniqueName: \"kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.691219 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.691877 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.691907 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.714276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc95g\" (UniqueName: \"kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:58 crc kubenswrapper[4754]: I1011 03:15:58.810773 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:15:59 crc kubenswrapper[4754]: I1011 03:15:59.291135 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs"] Oct 11 03:15:59 crc kubenswrapper[4754]: W1011 03:15:59.303701 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f472e73_bd23_4ff5_97de_14d0596c1b0c.slice/crio-382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a WatchSource:0}: Error finding container 382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a: Status 404 returned error can't find the container with id 382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.047050 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerID="642db9de6ae362cc9d6aab63d70bad5b0cc8976e8bb34998559e354d2e421d00" exitCode=0 Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.047171 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" event={"ID":"5f472e73-bd23-4ff5-97de-14d0596c1b0c","Type":"ContainerDied","Data":"642db9de6ae362cc9d6aab63d70bad5b0cc8976e8bb34998559e354d2e421d00"} Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.047224 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" event={"ID":"5f472e73-bd23-4ff5-97de-14d0596c1b0c","Type":"ContainerStarted","Data":"382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a"} Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.736748 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.736862 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.736941 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.737866 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:16:00 crc kubenswrapper[4754]: I1011 03:16:00.737996 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec" gracePeriod=600 Oct 11 03:16:01 crc kubenswrapper[4754]: I1011 03:16:01.060207 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec" exitCode=0 Oct 11 03:16:01 crc kubenswrapper[4754]: I1011 03:16:01.060289 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec"} Oct 11 03:16:01 crc kubenswrapper[4754]: I1011 03:16:01.060911 4754 scope.go:117] "RemoveContainer" containerID="a69d73945d577e5d4dc4c342b9ebc7b154e122e189f5f2d1630882b5734758fd" Oct 11 03:16:02 crc kubenswrapper[4754]: I1011 03:16:02.071827 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerID="e2ab5fad2c524248d77e13660de27269d8494bca923651e0199690c741a3ef58" exitCode=0 Oct 11 03:16:02 crc kubenswrapper[4754]: I1011 03:16:02.071927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" event={"ID":"5f472e73-bd23-4ff5-97de-14d0596c1b0c","Type":"ContainerDied","Data":"e2ab5fad2c524248d77e13660de27269d8494bca923651e0199690c741a3ef58"} Oct 11 03:16:02 crc kubenswrapper[4754]: I1011 03:16:02.075886 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac"} Oct 11 03:16:03 crc kubenswrapper[4754]: I1011 03:16:03.090192 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerID="d885396f837842c31f1336c6b4c38f2fe841d1e5321bac91b0eb95e58ae07dab" exitCode=0 Oct 11 03:16:03 crc kubenswrapper[4754]: I1011 03:16:03.095057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" event={"ID":"5f472e73-bd23-4ff5-97de-14d0596c1b0c","Type":"ContainerDied","Data":"d885396f837842c31f1336c6b4c38f2fe841d1e5321bac91b0eb95e58ae07dab"} Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.449613 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.579481 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc95g\" (UniqueName: \"kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g\") pod \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.579574 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle\") pod \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.580351 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util\") pod \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\" (UID: \"5f472e73-bd23-4ff5-97de-14d0596c1b0c\") " Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.580801 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle" (OuterVolumeSpecName: "bundle") pod "5f472e73-bd23-4ff5-97de-14d0596c1b0c" (UID: "5f472e73-bd23-4ff5-97de-14d0596c1b0c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.590200 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g" (OuterVolumeSpecName: "kube-api-access-vc95g") pod "5f472e73-bd23-4ff5-97de-14d0596c1b0c" (UID: "5f472e73-bd23-4ff5-97de-14d0596c1b0c"). InnerVolumeSpecName "kube-api-access-vc95g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.594331 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util" (OuterVolumeSpecName: "util") pod "5f472e73-bd23-4ff5-97de-14d0596c1b0c" (UID: "5f472e73-bd23-4ff5-97de-14d0596c1b0c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.681664 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.681715 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f472e73-bd23-4ff5-97de-14d0596c1b0c-util\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:04 crc kubenswrapper[4754]: I1011 03:16:04.681732 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc95g\" (UniqueName: \"kubernetes.io/projected/5f472e73-bd23-4ff5-97de-14d0596c1b0c-kube-api-access-vc95g\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:05 crc kubenswrapper[4754]: I1011 03:16:05.127311 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" event={"ID":"5f472e73-bd23-4ff5-97de-14d0596c1b0c","Type":"ContainerDied","Data":"382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a"} Oct 11 03:16:05 crc kubenswrapper[4754]: I1011 03:16:05.127830 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="382fd4f4b7d0b684aee4002e4ba15f65d2a56d6a30be94d78ff68b9c075df02a" Oct 11 03:16:05 crc kubenswrapper[4754]: I1011 03:16:05.127452 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.163277 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-tvh84"] Oct 11 03:16:06 crc kubenswrapper[4754]: E1011 03:16:06.164060 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="pull" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.164080 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="pull" Oct 11 03:16:06 crc kubenswrapper[4754]: E1011 03:16:06.164100 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="util" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.164108 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="util" Oct 11 03:16:06 crc kubenswrapper[4754]: E1011 03:16:06.164129 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="extract" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.164138 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="extract" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.164271 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f472e73-bd23-4ff5-97de-14d0596c1b0c" containerName="extract" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.164892 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.188988 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.189272 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zl5x7" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.189405 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.192551 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-tvh84"] Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.207948 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6s69\" (UniqueName: \"kubernetes.io/projected/dd5cec53-cb56-472d-94c9-af438d539f17-kube-api-access-m6s69\") pod \"nmstate-operator-858ddd8f98-tvh84\" (UID: \"dd5cec53-cb56-472d-94c9-af438d539f17\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.309488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6s69\" (UniqueName: \"kubernetes.io/projected/dd5cec53-cb56-472d-94c9-af438d539f17-kube-api-access-m6s69\") pod \"nmstate-operator-858ddd8f98-tvh84\" (UID: \"dd5cec53-cb56-472d-94c9-af438d539f17\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.332529 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6s69\" (UniqueName: \"kubernetes.io/projected/dd5cec53-cb56-472d-94c9-af438d539f17-kube-api-access-m6s69\") pod \"nmstate-operator-858ddd8f98-tvh84\" (UID: \"dd5cec53-cb56-472d-94c9-af438d539f17\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.524391 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" Oct 11 03:16:06 crc kubenswrapper[4754]: I1011 03:16:06.768439 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-tvh84"] Oct 11 03:16:07 crc kubenswrapper[4754]: I1011 03:16:07.139618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" event={"ID":"dd5cec53-cb56-472d-94c9-af438d539f17","Type":"ContainerStarted","Data":"547a69364192f4002275a66cdac9a040d66c01e585089578b82411e6bedca17d"} Oct 11 03:16:10 crc kubenswrapper[4754]: I1011 03:16:10.162081 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" event={"ID":"dd5cec53-cb56-472d-94c9-af438d539f17","Type":"ContainerStarted","Data":"8c5a4ac28fed05d873e0aabeabfae2a1ecc8759b201dea42dfa8fa45880cf13a"} Oct 11 03:16:10 crc kubenswrapper[4754]: I1011 03:16:10.193482 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-tvh84" podStartSLOduration=1.075806973 podStartE2EDuration="4.193457275s" podCreationTimestamp="2025-10-11 03:16:06 +0000 UTC" firstStartedPulling="2025-10-11 03:16:06.787331649 +0000 UTC m=+614.346276434" lastFinishedPulling="2025-10-11 03:16:09.904981941 +0000 UTC m=+617.463926736" observedRunningTime="2025-10-11 03:16:10.186802043 +0000 UTC m=+617.745746858" watchObservedRunningTime="2025-10-11 03:16:10.193457275 +0000 UTC m=+617.752402070" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.164177 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.165453 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.172188 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-cdwjw" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.178253 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.179496 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.182870 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.187922 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.194772 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7qn6k"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.195691 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.201883 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283596 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k874p\" (UniqueName: \"kubernetes.io/projected/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-kube-api-access-k874p\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/446f8e6b-697d-4bfa-8464-e8b4decf2a33-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283677 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-nmstate-lock\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-dbus-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-ovs-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dkzn\" (UniqueName: \"kubernetes.io/projected/446f8e6b-697d-4bfa-8464-e8b4decf2a33-kube-api-access-6dkzn\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.283847 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mspqt\" (UniqueName: \"kubernetes.io/projected/d39cf2da-c123-476f-abc2-acfe1da12347-kube-api-access-mspqt\") pod \"nmstate-metrics-fdff9cb8d-2s4dm\" (UID: \"d39cf2da-c123-476f-abc2-acfe1da12347\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.322592 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.323757 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.326098 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.326469 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.332438 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9r2xs" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.340428 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.384865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-ovs-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.384932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmxpw\" (UniqueName: \"kubernetes.io/projected/bb182ceb-1372-432d-9809-c67c03eb4091-kube-api-access-zmxpw\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.384991 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dkzn\" (UniqueName: \"kubernetes.io/projected/446f8e6b-697d-4bfa-8464-e8b4decf2a33-kube-api-access-6dkzn\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb182ceb-1372-432d-9809-c67c03eb4091-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mspqt\" (UniqueName: \"kubernetes.io/projected/d39cf2da-c123-476f-abc2-acfe1da12347-kube-api-access-mspqt\") pod \"nmstate-metrics-fdff9cb8d-2s4dm\" (UID: \"d39cf2da-c123-476f-abc2-acfe1da12347\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385050 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-ovs-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k874p\" (UniqueName: \"kubernetes.io/projected/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-kube-api-access-k874p\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385211 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/446f8e6b-697d-4bfa-8464-e8b4decf2a33-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385261 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-nmstate-lock\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-dbus-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385407 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385593 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-nmstate-lock\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.385936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-dbus-socket\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.400923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/446f8e6b-697d-4bfa-8464-e8b4decf2a33-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.404597 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mspqt\" (UniqueName: \"kubernetes.io/projected/d39cf2da-c123-476f-abc2-acfe1da12347-kube-api-access-mspqt\") pod \"nmstate-metrics-fdff9cb8d-2s4dm\" (UID: \"d39cf2da-c123-476f-abc2-acfe1da12347\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.409247 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dkzn\" (UniqueName: \"kubernetes.io/projected/446f8e6b-697d-4bfa-8464-e8b4decf2a33-kube-api-access-6dkzn\") pod \"nmstate-webhook-6cdbc54649-b4vlh\" (UID: \"446f8e6b-697d-4bfa-8464-e8b4decf2a33\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.410880 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k874p\" (UniqueName: \"kubernetes.io/projected/2a01ddcb-c062-4d74-91e7-a6793e5b79cb-kube-api-access-k874p\") pod \"nmstate-handler-7qn6k\" (UID: \"2a01ddcb-c062-4d74-91e7-a6793e5b79cb\") " pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.485313 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.486713 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmxpw\" (UniqueName: \"kubernetes.io/projected/bb182ceb-1372-432d-9809-c67c03eb4091-kube-api-access-zmxpw\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.486767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb182ceb-1372-432d-9809-c67c03eb4091-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.486841 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: E1011 03:16:11.487050 4754 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 11 03:16:11 crc kubenswrapper[4754]: E1011 03:16:11.487131 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert podName:bb182ceb-1372-432d-9809-c67c03eb4091 nodeName:}" failed. No retries permitted until 2025-10-11 03:16:11.987100061 +0000 UTC m=+619.546044846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-txjs2" (UID: "bb182ceb-1372-432d-9809-c67c03eb4091") : secret "plugin-serving-cert" not found Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.488349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb182ceb-1372-432d-9809-c67c03eb4091-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.506919 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmxpw\" (UniqueName: \"kubernetes.io/projected/bb182ceb-1372-432d-9809-c67c03eb4091-kube-api-access-zmxpw\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.513154 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.524689 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86d4db95d4-2ntrt"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.527780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.530267 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.555301 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86d4db95d4-2ntrt"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-service-ca\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-oauth-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588354 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588441 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-trusted-ca-bundle\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588481 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-oauth-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvzsx\" (UniqueName: \"kubernetes.io/projected/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-kube-api-access-wvzsx\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.588580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691567 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvzsx\" (UniqueName: \"kubernetes.io/projected/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-kube-api-access-wvzsx\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691629 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691665 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-service-ca\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691683 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-oauth-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-trusted-ca-bundle\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.691798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-oauth-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.693058 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-service-ca\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.693236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.693373 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-oauth-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.694009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-trusted-ca-bundle\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.695206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-serving-cert\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.698512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-console-oauth-config\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.708479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvzsx\" (UniqueName: \"kubernetes.io/projected/92bd5a14-0ae3-4236-afbf-cac0e9aea33e-kube-api-access-wvzsx\") pod \"console-86d4db95d4-2ntrt\" (UID: \"92bd5a14-0ae3-4236-afbf-cac0e9aea33e\") " pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.757897 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm"] Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.795397 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh"] Oct 11 03:16:11 crc kubenswrapper[4754]: W1011 03:16:11.805483 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod446f8e6b_697d_4bfa_8464_e8b4decf2a33.slice/crio-9fe90a55115624fe773ac3cdc677f3fc91969dcba61e1ec36fac3f200195de48 WatchSource:0}: Error finding container 9fe90a55115624fe773ac3cdc677f3fc91969dcba61e1ec36fac3f200195de48: Status 404 returned error can't find the container with id 9fe90a55115624fe773ac3cdc677f3fc91969dcba61e1ec36fac3f200195de48 Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.882378 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:11 crc kubenswrapper[4754]: I1011 03:16:11.997731 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.004403 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb182ceb-1372-432d-9809-c67c03eb4091-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-txjs2\" (UID: \"bb182ceb-1372-432d-9809-c67c03eb4091\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.175162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" event={"ID":"446f8e6b-697d-4bfa-8464-e8b4decf2a33","Type":"ContainerStarted","Data":"9fe90a55115624fe773ac3cdc677f3fc91969dcba61e1ec36fac3f200195de48"} Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.177127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" event={"ID":"d39cf2da-c123-476f-abc2-acfe1da12347","Type":"ContainerStarted","Data":"82a528edc095c85701fdc064d7f19361efbcb120adb0273cba825b3645d54bab"} Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.178578 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7qn6k" event={"ID":"2a01ddcb-c062-4d74-91e7-a6793e5b79cb","Type":"ContainerStarted","Data":"36af9719e3e19b48664e48c570cc2728ef8e1b50f39a21edcc067deed0553c4a"} Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.240531 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.366862 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86d4db95d4-2ntrt"] Oct 11 03:16:12 crc kubenswrapper[4754]: I1011 03:16:12.501733 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2"] Oct 11 03:16:12 crc kubenswrapper[4754]: W1011 03:16:12.518207 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb182ceb_1372_432d_9809_c67c03eb4091.slice/crio-0aa7f36731975fda39ac6371393f0923021465bf6693731d4dde891d07b53afe WatchSource:0}: Error finding container 0aa7f36731975fda39ac6371393f0923021465bf6693731d4dde891d07b53afe: Status 404 returned error can't find the container with id 0aa7f36731975fda39ac6371393f0923021465bf6693731d4dde891d07b53afe Oct 11 03:16:13 crc kubenswrapper[4754]: I1011 03:16:13.189410 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d4db95d4-2ntrt" event={"ID":"92bd5a14-0ae3-4236-afbf-cac0e9aea33e","Type":"ContainerStarted","Data":"a66906405eb659bb97821b38194a86b4a894418f1a99d536d8b27a240513a997"} Oct 11 03:16:13 crc kubenswrapper[4754]: I1011 03:16:13.189760 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d4db95d4-2ntrt" event={"ID":"92bd5a14-0ae3-4236-afbf-cac0e9aea33e","Type":"ContainerStarted","Data":"1cafd3d94fd83a8b7332b0c74ad44c6febb9075218f5854e6f41a459fe934455"} Oct 11 03:16:13 crc kubenswrapper[4754]: I1011 03:16:13.193869 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" event={"ID":"bb182ceb-1372-432d-9809-c67c03eb4091","Type":"ContainerStarted","Data":"0aa7f36731975fda39ac6371393f0923021465bf6693731d4dde891d07b53afe"} Oct 11 03:16:13 crc kubenswrapper[4754]: I1011 03:16:13.212581 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86d4db95d4-2ntrt" podStartSLOduration=2.212557132 podStartE2EDuration="2.212557132s" podCreationTimestamp="2025-10-11 03:16:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:16:13.209084662 +0000 UTC m=+620.768029457" watchObservedRunningTime="2025-10-11 03:16:13.212557132 +0000 UTC m=+620.771501917" Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.219535 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7qn6k" event={"ID":"2a01ddcb-c062-4d74-91e7-a6793e5b79cb","Type":"ContainerStarted","Data":"47b93c818ae2421900acb9ccce4f4b050b3bea9163c1db00ec69a1538299172f"} Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.220251 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.224307 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" event={"ID":"446f8e6b-697d-4bfa-8464-e8b4decf2a33","Type":"ContainerStarted","Data":"be4337efcaadcfd9416902f4d8576795fad51349d2253583107c6e80c4aede1a"} Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.224348 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.225818 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" event={"ID":"d39cf2da-c123-476f-abc2-acfe1da12347","Type":"ContainerStarted","Data":"9f3eb45fe169fbf368a5b5d3703d65e23366caadd91c3cb1b3bbce18063ff272"} Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.275258 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7qn6k" podStartSLOduration=1.746521746 podStartE2EDuration="4.275232791s" podCreationTimestamp="2025-10-11 03:16:11 +0000 UTC" firstStartedPulling="2025-10-11 03:16:11.563046671 +0000 UTC m=+619.121991456" lastFinishedPulling="2025-10-11 03:16:14.091757716 +0000 UTC m=+621.650702501" observedRunningTime="2025-10-11 03:16:15.240422473 +0000 UTC m=+622.799367258" watchObservedRunningTime="2025-10-11 03:16:15.275232791 +0000 UTC m=+622.834177566" Oct 11 03:16:15 crc kubenswrapper[4754]: I1011 03:16:15.278594 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" podStartSLOduration=1.909202898 podStartE2EDuration="4.278577118s" podCreationTimestamp="2025-10-11 03:16:11 +0000 UTC" firstStartedPulling="2025-10-11 03:16:11.80748672 +0000 UTC m=+619.366431505" lastFinishedPulling="2025-10-11 03:16:14.17686093 +0000 UTC m=+621.735805725" observedRunningTime="2025-10-11 03:16:15.269251748 +0000 UTC m=+622.828196543" watchObservedRunningTime="2025-10-11 03:16:15.278577118 +0000 UTC m=+622.837521903" Oct 11 03:16:16 crc kubenswrapper[4754]: I1011 03:16:16.234122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" event={"ID":"bb182ceb-1372-432d-9809-c67c03eb4091","Type":"ContainerStarted","Data":"df1de3a818fc2c48b6933d1193a26c401d6434b0672b81397f834ac9b7169e16"} Oct 11 03:16:16 crc kubenswrapper[4754]: I1011 03:16:16.256149 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-txjs2" podStartSLOduration=2.637901722 podStartE2EDuration="5.256121899s" podCreationTimestamp="2025-10-11 03:16:11 +0000 UTC" firstStartedPulling="2025-10-11 03:16:12.520382117 +0000 UTC m=+620.079326902" lastFinishedPulling="2025-10-11 03:16:15.138602294 +0000 UTC m=+622.697547079" observedRunningTime="2025-10-11 03:16:16.24819745 +0000 UTC m=+623.807142265" watchObservedRunningTime="2025-10-11 03:16:16.256121899 +0000 UTC m=+623.815066684" Oct 11 03:16:17 crc kubenswrapper[4754]: I1011 03:16:17.241146 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" event={"ID":"d39cf2da-c123-476f-abc2-acfe1da12347","Type":"ContainerStarted","Data":"3ccbd857877e2a33c575eb367bc0a8680cabc587187d4cf2351f7afa51fd6445"} Oct 11 03:16:17 crc kubenswrapper[4754]: I1011 03:16:17.260279 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2s4dm" podStartSLOduration=1.375365397 podStartE2EDuration="6.260257509s" podCreationTimestamp="2025-10-11 03:16:11 +0000 UTC" firstStartedPulling="2025-10-11 03:16:11.783132115 +0000 UTC m=+619.342076900" lastFinishedPulling="2025-10-11 03:16:16.668024237 +0000 UTC m=+624.226969012" observedRunningTime="2025-10-11 03:16:17.257123819 +0000 UTC m=+624.816068664" watchObservedRunningTime="2025-10-11 03:16:17.260257509 +0000 UTC m=+624.819202294" Oct 11 03:16:21 crc kubenswrapper[4754]: I1011 03:16:21.571198 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7qn6k" Oct 11 03:16:21 crc kubenswrapper[4754]: I1011 03:16:21.883150 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:21 crc kubenswrapper[4754]: I1011 03:16:21.883222 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:21 crc kubenswrapper[4754]: I1011 03:16:21.887931 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:22 crc kubenswrapper[4754]: I1011 03:16:22.280611 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86d4db95d4-2ntrt" Oct 11 03:16:22 crc kubenswrapper[4754]: I1011 03:16:22.337626 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:16:31 crc kubenswrapper[4754]: I1011 03:16:31.520569 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b4vlh" Oct 11 03:16:46 crc kubenswrapper[4754]: I1011 03:16:46.959641 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr"] Oct 11 03:16:46 crc kubenswrapper[4754]: I1011 03:16:46.961501 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:46 crc kubenswrapper[4754]: I1011 03:16:46.964722 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 11 03:16:46 crc kubenswrapper[4754]: I1011 03:16:46.971779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr"] Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.071162 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.071214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.071438 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r47hz\" (UniqueName: \"kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.172759 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.172808 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.172850 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r47hz\" (UniqueName: \"kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.173507 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.173602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.197774 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r47hz\" (UniqueName: \"kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.293155 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.392496 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-m25hk" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" containerID="cri-o://f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037" gracePeriod=15 Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.535703 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr"] Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.728005 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m25hk_36eab906-00ee-4a62-b0c1-85f3daccb0d8/console/0.log" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.728491 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.882414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.882551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.884180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.884285 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.884319 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.884349 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hnzg\" (UniqueName: \"kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.884383 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config\") pod \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\" (UID: \"36eab906-00ee-4a62-b0c1-85f3daccb0d8\") " Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.885134 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.885164 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca" (OuterVolumeSpecName: "service-ca") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.885210 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.885658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config" (OuterVolumeSpecName: "console-config") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.891033 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg" (OuterVolumeSpecName: "kube-api-access-5hnzg") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "kube-api-access-5hnzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.891033 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.891256 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "36eab906-00ee-4a62-b0c1-85f3daccb0d8" (UID: "36eab906-00ee-4a62-b0c1-85f3daccb0d8"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985854 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985897 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985910 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985924 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985937 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985949 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hnzg\" (UniqueName: \"kubernetes.io/projected/36eab906-00ee-4a62-b0c1-85f3daccb0d8-kube-api-access-5hnzg\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:47 crc kubenswrapper[4754]: I1011 03:16:47.985980 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36eab906-00ee-4a62-b0c1-85f3daccb0d8-console-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.478015 4754 generic.go:334] "Generic (PLEG): container finished" podID="25829584-c627-4829-933c-fa2420a920b7" containerID="25075cb31252f158952b5f302f3c4d40a7b364c9455f319cf8f3e3f77afe04f9" exitCode=0 Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.478108 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" event={"ID":"25829584-c627-4829-933c-fa2420a920b7","Type":"ContainerDied","Data":"25075cb31252f158952b5f302f3c4d40a7b364c9455f319cf8f3e3f77afe04f9"} Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.478471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" event={"ID":"25829584-c627-4829-933c-fa2420a920b7","Type":"ContainerStarted","Data":"c0783b09f30aaf4096704e7b91316c6994cbf81ede89dd95aa479ab14e20962a"} Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.480950 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m25hk_36eab906-00ee-4a62-b0c1-85f3daccb0d8/console/0.log" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.481015 4754 generic.go:334] "Generic (PLEG): container finished" podID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerID="f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037" exitCode=2 Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.481042 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m25hk" event={"ID":"36eab906-00ee-4a62-b0c1-85f3daccb0d8","Type":"ContainerDied","Data":"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037"} Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.481065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m25hk" event={"ID":"36eab906-00ee-4a62-b0c1-85f3daccb0d8","Type":"ContainerDied","Data":"21214b6adfcfa8e11872a43f61132659fd341b9ec5dc03d03ea66ecf94ad13cd"} Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.481082 4754 scope.go:117] "RemoveContainer" containerID="f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.481114 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m25hk" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.500252 4754 scope.go:117] "RemoveContainer" containerID="f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037" Oct 11 03:16:48 crc kubenswrapper[4754]: E1011 03:16:48.500734 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037\": container with ID starting with f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037 not found: ID does not exist" containerID="f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.500775 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037"} err="failed to get container status \"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037\": rpc error: code = NotFound desc = could not find container \"f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037\": container with ID starting with f966178ffe18c59974220b63a20549847eedec6766d972c6e52aef3e980d2037 not found: ID does not exist" Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.526414 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:16:48 crc kubenswrapper[4754]: I1011 03:16:48.545628 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-m25hk"] Oct 11 03:16:49 crc kubenswrapper[4754]: I1011 03:16:49.100152 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" path="/var/lib/kubelet/pods/36eab906-00ee-4a62-b0c1-85f3daccb0d8/volumes" Oct 11 03:16:50 crc kubenswrapper[4754]: I1011 03:16:50.503617 4754 generic.go:334] "Generic (PLEG): container finished" podID="25829584-c627-4829-933c-fa2420a920b7" containerID="b8a379f7fb9330b489535fe105599b02389dca444c7ed1675a084a4fb61a458f" exitCode=0 Oct 11 03:16:50 crc kubenswrapper[4754]: I1011 03:16:50.503741 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" event={"ID":"25829584-c627-4829-933c-fa2420a920b7","Type":"ContainerDied","Data":"b8a379f7fb9330b489535fe105599b02389dca444c7ed1675a084a4fb61a458f"} Oct 11 03:16:51 crc kubenswrapper[4754]: I1011 03:16:51.516604 4754 generic.go:334] "Generic (PLEG): container finished" podID="25829584-c627-4829-933c-fa2420a920b7" containerID="34ea09992e2f8c9e848f55d2fcfea3f26b4943f0b61345dd6e088198f47cb38f" exitCode=0 Oct 11 03:16:51 crc kubenswrapper[4754]: I1011 03:16:51.516729 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" event={"ID":"25829584-c627-4829-933c-fa2420a920b7","Type":"ContainerDied","Data":"34ea09992e2f8c9e848f55d2fcfea3f26b4943f0b61345dd6e088198f47cb38f"} Oct 11 03:16:52 crc kubenswrapper[4754]: I1011 03:16:52.900241 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.079277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle\") pod \"25829584-c627-4829-933c-fa2420a920b7\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.079373 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r47hz\" (UniqueName: \"kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz\") pod \"25829584-c627-4829-933c-fa2420a920b7\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.079425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util\") pod \"25829584-c627-4829-933c-fa2420a920b7\" (UID: \"25829584-c627-4829-933c-fa2420a920b7\") " Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.080722 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle" (OuterVolumeSpecName: "bundle") pod "25829584-c627-4829-933c-fa2420a920b7" (UID: "25829584-c627-4829-933c-fa2420a920b7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.081026 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.085825 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz" (OuterVolumeSpecName: "kube-api-access-r47hz") pod "25829584-c627-4829-933c-fa2420a920b7" (UID: "25829584-c627-4829-933c-fa2420a920b7"). InnerVolumeSpecName "kube-api-access-r47hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.178486 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util" (OuterVolumeSpecName: "util") pod "25829584-c627-4829-933c-fa2420a920b7" (UID: "25829584-c627-4829-933c-fa2420a920b7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.181793 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25829584-c627-4829-933c-fa2420a920b7-util\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.181839 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r47hz\" (UniqueName: \"kubernetes.io/projected/25829584-c627-4829-933c-fa2420a920b7-kube-api-access-r47hz\") on node \"crc\" DevicePath \"\"" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.532602 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" event={"ID":"25829584-c627-4829-933c-fa2420a920b7","Type":"ContainerDied","Data":"c0783b09f30aaf4096704e7b91316c6994cbf81ede89dd95aa479ab14e20962a"} Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.532670 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0783b09f30aaf4096704e7b91316c6994cbf81ede89dd95aa479ab14e20962a" Oct 11 03:16:53 crc kubenswrapper[4754]: I1011 03:16:53.533043 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.673991 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm"] Oct 11 03:17:02 crc kubenswrapper[4754]: E1011 03:17:02.674958 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679078 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" Oct 11 03:17:02 crc kubenswrapper[4754]: E1011 03:17:02.679138 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="util" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679147 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="util" Oct 11 03:17:02 crc kubenswrapper[4754]: E1011 03:17:02.679158 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="pull" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679166 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="pull" Oct 11 03:17:02 crc kubenswrapper[4754]: E1011 03:17:02.679183 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="extract" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679206 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="extract" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679420 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="36eab906-00ee-4a62-b0c1-85f3daccb0d8" containerName="console" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.679454 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="25829584-c627-4829-933c-fa2420a920b7" containerName="extract" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.680162 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.697702 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.697909 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.698037 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.698591 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pxcgh" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.698653 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.740629 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-webhook-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.740980 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2fnt\" (UniqueName: \"kubernetes.io/projected/37697b5b-13b1-40b7-96e6-4e5e21499bec-kube-api-access-f2fnt\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.741080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-apiservice-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.744845 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm"] Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.841911 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-webhook-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.841987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2fnt\" (UniqueName: \"kubernetes.io/projected/37697b5b-13b1-40b7-96e6-4e5e21499bec-kube-api-access-f2fnt\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.842007 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-apiservice-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.849530 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-apiservice-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.849587 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37697b5b-13b1-40b7-96e6-4e5e21499bec-webhook-cert\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:02 crc kubenswrapper[4754]: I1011 03:17:02.862659 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2fnt\" (UniqueName: \"kubernetes.io/projected/37697b5b-13b1-40b7-96e6-4e5e21499bec-kube-api-access-f2fnt\") pod \"metallb-operator-controller-manager-67f8c59cbf-xd9pm\" (UID: \"37697b5b-13b1-40b7-96e6-4e5e21499bec\") " pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.048956 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.054726 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4"] Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.055509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.057151 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.057181 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6h5h8" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.073979 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.074061 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4"] Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.245954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgrzf\" (UniqueName: \"kubernetes.io/projected/ebdfdd95-930a-43be-b2c4-1981a60df09e-kube-api-access-cgrzf\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.246033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-webhook-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.246083 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-apiservice-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.282980 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm"] Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.347618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-apiservice-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.348163 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgrzf\" (UniqueName: \"kubernetes.io/projected/ebdfdd95-930a-43be-b2c4-1981a60df09e-kube-api-access-cgrzf\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.348197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-webhook-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.355651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-webhook-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.355651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebdfdd95-930a-43be-b2c4-1981a60df09e-apiservice-cert\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.370040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgrzf\" (UniqueName: \"kubernetes.io/projected/ebdfdd95-930a-43be-b2c4-1981a60df09e-kube-api-access-cgrzf\") pod \"metallb-operator-webhook-server-84c7b6d8cc-b6cv4\" (UID: \"ebdfdd95-930a-43be-b2c4-1981a60df09e\") " pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.422270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.632004 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4"] Oct 11 03:17:03 crc kubenswrapper[4754]: I1011 03:17:03.641208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" event={"ID":"37697b5b-13b1-40b7-96e6-4e5e21499bec","Type":"ContainerStarted","Data":"bdf982ad096b3612a7589cc2f9e5f08c059a53f4d1f1df9d5a990106589ac835"} Oct 11 03:17:03 crc kubenswrapper[4754]: W1011 03:17:03.641608 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebdfdd95_930a_43be_b2c4_1981a60df09e.slice/crio-3ed1b3ba0495c0a48e58d0cac04a50aeee485fe78da1cc5f9e6c3aa4bdc31de0 WatchSource:0}: Error finding container 3ed1b3ba0495c0a48e58d0cac04a50aeee485fe78da1cc5f9e6c3aa4bdc31de0: Status 404 returned error can't find the container with id 3ed1b3ba0495c0a48e58d0cac04a50aeee485fe78da1cc5f9e6c3aa4bdc31de0 Oct 11 03:17:04 crc kubenswrapper[4754]: I1011 03:17:04.647780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" event={"ID":"ebdfdd95-930a-43be-b2c4-1981a60df09e","Type":"ContainerStarted","Data":"3ed1b3ba0495c0a48e58d0cac04a50aeee485fe78da1cc5f9e6c3aa4bdc31de0"} Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.677872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" event={"ID":"37697b5b-13b1-40b7-96e6-4e5e21499bec","Type":"ContainerStarted","Data":"c94c9979994ce7d8f55bcd73ead67e54f870bbb70369ba2f25aaf1411441960e"} Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.678523 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.679689 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" event={"ID":"ebdfdd95-930a-43be-b2c4-1981a60df09e","Type":"ContainerStarted","Data":"2c0d43e1ea9ee406ab2a6b913c76b95b3eba3fe8ef86f718624c11bd621049fa"} Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.679842 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.700080 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" podStartSLOduration=2.214762228 podStartE2EDuration="7.700061513s" podCreationTimestamp="2025-10-11 03:17:02 +0000 UTC" firstStartedPulling="2025-10-11 03:17:03.292527781 +0000 UTC m=+670.851472566" lastFinishedPulling="2025-10-11 03:17:08.777827066 +0000 UTC m=+676.336771851" observedRunningTime="2025-10-11 03:17:09.697065448 +0000 UTC m=+677.256010253" watchObservedRunningTime="2025-10-11 03:17:09.700061513 +0000 UTC m=+677.259006308" Oct 11 03:17:09 crc kubenswrapper[4754]: I1011 03:17:09.718270 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" podStartSLOduration=1.571149331 podStartE2EDuration="6.718248347s" podCreationTimestamp="2025-10-11 03:17:03 +0000 UTC" firstStartedPulling="2025-10-11 03:17:03.644705356 +0000 UTC m=+671.203650141" lastFinishedPulling="2025-10-11 03:17:08.791804372 +0000 UTC m=+676.350749157" observedRunningTime="2025-10-11 03:17:09.714913643 +0000 UTC m=+677.273858438" watchObservedRunningTime="2025-10-11 03:17:09.718248347 +0000 UTC m=+677.277193132" Oct 11 03:17:23 crc kubenswrapper[4754]: I1011 03:17:23.429538 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-84c7b6d8cc-b6cv4" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.052367 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-67f8c59cbf-xd9pm" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.919921 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw"] Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.921139 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.924200 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-qb968" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.925190 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.931292 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw"] Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.941176 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wjqkr"] Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.943990 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.949510 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.950121 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.997896 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-49m5p"] Oct 11 03:17:43 crc kubenswrapper[4754]: I1011 03:17:43.999145 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.000928 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.001326 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.001574 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.001771 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-b9lgd" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.001788 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-9s44v"] Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.002995 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.004700 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.017343 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-9s44v"] Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116298 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics-certs\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-sockets\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116765 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-reloader\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skw72\" (UniqueName: \"kubernetes.io/projected/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-kube-api-access-skw72\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116805 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lskg\" (UniqueName: \"kubernetes.io/projected/02c9fc4b-bec8-4468-b120-08f4db18ed07-kube-api-access-7lskg\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116891 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65gvq\" (UniqueName: \"kubernetes.io/projected/01ff4e21-2909-48d9-a3e6-dae8528a1658-kube-api-access-65gvq\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01ff4e21-2909-48d9-a3e6-dae8528a1658-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116925 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-conf\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116945 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metallb-excludel2\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.116983 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-cert\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.117003 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metrics-certs\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.117057 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m5rj\" (UniqueName: \"kubernetes.io/projected/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-kube-api-access-9m5rj\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.117092 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-startup\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-cert\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metrics-certs\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m5rj\" (UniqueName: \"kubernetes.io/projected/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-kube-api-access-9m5rj\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-startup\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217855 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217893 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics-certs\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217909 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-sockets\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217926 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-reloader\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skw72\" (UniqueName: \"kubernetes.io/projected/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-kube-api-access-skw72\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.217991 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lskg\" (UniqueName: \"kubernetes.io/projected/02c9fc4b-bec8-4468-b120-08f4db18ed07-kube-api-access-7lskg\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65gvq\" (UniqueName: \"kubernetes.io/projected/01ff4e21-2909-48d9-a3e6-dae8528a1658-kube-api-access-65gvq\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218049 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01ff4e21-2909-48d9-a3e6-dae8528a1658-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-conf\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metallb-excludel2\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218433 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.218695 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metallb-excludel2\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.218738 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.218801 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist podName:70ca9b96-8096-4d8a-b1e6-f954eb7de59f nodeName:}" failed. No retries permitted until 2025-10-11 03:17:44.718778666 +0000 UTC m=+712.277723501 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist") pod "speaker-49m5p" (UID: "70ca9b96-8096-4d8a-b1e6-f954eb7de59f") : secret "metallb-memberlist" not found Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.218809 4754 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.218858 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs podName:fa5eef15-bdc7-497b-aae1-0a6d1f5d3885 nodeName:}" failed. No retries permitted until 2025-10-11 03:17:44.718840698 +0000 UTC m=+712.277785563 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs") pod "controller-68d546b9d8-9s44v" (UID: "fa5eef15-bdc7-497b-aae1-0a6d1f5d3885") : secret "controller-certs-secret" not found Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.219227 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-reloader\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.219228 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-sockets\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.219344 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-conf\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.220279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/02c9fc4b-bec8-4468-b120-08f4db18ed07-frr-startup\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.223346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-metrics-certs\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.227884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-cert\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.243024 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m5rj\" (UniqueName: \"kubernetes.io/projected/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-kube-api-access-9m5rj\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.257029 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skw72\" (UniqueName: \"kubernetes.io/projected/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-kube-api-access-skw72\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.263317 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01ff4e21-2909-48d9-a3e6-dae8528a1658-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.273512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02c9fc4b-bec8-4468-b120-08f4db18ed07-metrics-certs\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.279707 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lskg\" (UniqueName: \"kubernetes.io/projected/02c9fc4b-bec8-4468-b120-08f4db18ed07-kube-api-access-7lskg\") pod \"frr-k8s-wjqkr\" (UID: \"02c9fc4b-bec8-4468-b120-08f4db18ed07\") " pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.292768 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65gvq\" (UniqueName: \"kubernetes.io/projected/01ff4e21-2909-48d9-a3e6-dae8528a1658-kube-api-access-65gvq\") pod \"frr-k8s-webhook-server-64bf5d555-6rwvw\" (UID: \"01ff4e21-2909-48d9-a3e6-dae8528a1658\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.539480 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.559996 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.724092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.724582 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.725126 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 11 03:17:44 crc kubenswrapper[4754]: E1011 03:17:44.725191 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist podName:70ca9b96-8096-4d8a-b1e6-f954eb7de59f nodeName:}" failed. No retries permitted until 2025-10-11 03:17:45.725174006 +0000 UTC m=+713.284118791 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist") pod "speaker-49m5p" (UID: "70ca9b96-8096-4d8a-b1e6-f954eb7de59f") : secret "metallb-memberlist" not found Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.730214 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa5eef15-bdc7-497b-aae1-0a6d1f5d3885-metrics-certs\") pod \"controller-68d546b9d8-9s44v\" (UID: \"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885\") " pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.915238 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"a881c678d3c6b21b07995c6d9e99712349d2cb90a0b1d67fcf9c7efe4c7448dc"} Oct 11 03:17:44 crc kubenswrapper[4754]: I1011 03:17:44.927630 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.023485 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw"] Oct 11 03:17:45 crc kubenswrapper[4754]: W1011 03:17:45.037541 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ff4e21_2909_48d9_a3e6_dae8528a1658.slice/crio-464906826a2a9adbf4db3e1c5f1f49509ee6ac4e67df9c314a6aeab84adbadd4 WatchSource:0}: Error finding container 464906826a2a9adbf4db3e1c5f1f49509ee6ac4e67df9c314a6aeab84adbadd4: Status 404 returned error can't find the container with id 464906826a2a9adbf4db3e1c5f1f49509ee6ac4e67df9c314a6aeab84adbadd4 Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.133667 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-9s44v"] Oct 11 03:17:45 crc kubenswrapper[4754]: W1011 03:17:45.143869 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa5eef15_bdc7_497b_aae1_0a6d1f5d3885.slice/crio-e9674e6a0618302dd3493df31d2b7579f3b8a2ac3870777e0526c442c896fd0a WatchSource:0}: Error finding container e9674e6a0618302dd3493df31d2b7579f3b8a2ac3870777e0526c442c896fd0a: Status 404 returned error can't find the container with id e9674e6a0618302dd3493df31d2b7579f3b8a2ac3870777e0526c442c896fd0a Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.740425 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.752813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/70ca9b96-8096-4d8a-b1e6-f954eb7de59f-memberlist\") pod \"speaker-49m5p\" (UID: \"70ca9b96-8096-4d8a-b1e6-f954eb7de59f\") " pod="metallb-system/speaker-49m5p" Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.816390 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-49m5p" Oct 11 03:17:45 crc kubenswrapper[4754]: W1011 03:17:45.850392 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70ca9b96_8096_4d8a_b1e6_f954eb7de59f.slice/crio-7f4337c70e53ba82bf010afecd0a557cad5ce50f4018c334ec9b4e4a7aaa0d96 WatchSource:0}: Error finding container 7f4337c70e53ba82bf010afecd0a557cad5ce50f4018c334ec9b4e4a7aaa0d96: Status 404 returned error can't find the container with id 7f4337c70e53ba82bf010afecd0a557cad5ce50f4018c334ec9b4e4a7aaa0d96 Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.930245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" event={"ID":"01ff4e21-2909-48d9-a3e6-dae8528a1658","Type":"ContainerStarted","Data":"464906826a2a9adbf4db3e1c5f1f49509ee6ac4e67df9c314a6aeab84adbadd4"} Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.931775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49m5p" event={"ID":"70ca9b96-8096-4d8a-b1e6-f954eb7de59f","Type":"ContainerStarted","Data":"7f4337c70e53ba82bf010afecd0a557cad5ce50f4018c334ec9b4e4a7aaa0d96"} Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.933934 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-9s44v" event={"ID":"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885","Type":"ContainerStarted","Data":"9b51d93bc02af27ea2baa9de83c700f2de74d61453a0cfa8cdcd3b3b3a7aaff9"} Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.933977 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-9s44v" event={"ID":"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885","Type":"ContainerStarted","Data":"0ba82dc3436c60202fcca8951ef97f7afd0ff3d776f29d2841e64ac8a4227476"} Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.933992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-9s44v" event={"ID":"fa5eef15-bdc7-497b-aae1-0a6d1f5d3885","Type":"ContainerStarted","Data":"e9674e6a0618302dd3493df31d2b7579f3b8a2ac3870777e0526c442c896fd0a"} Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.934088 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:17:45 crc kubenswrapper[4754]: I1011 03:17:45.954200 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-9s44v" podStartSLOduration=2.954177761 podStartE2EDuration="2.954177761s" podCreationTimestamp="2025-10-11 03:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:17:45.947662377 +0000 UTC m=+713.506607172" watchObservedRunningTime="2025-10-11 03:17:45.954177761 +0000 UTC m=+713.513122536" Oct 11 03:17:46 crc kubenswrapper[4754]: I1011 03:17:46.943912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49m5p" event={"ID":"70ca9b96-8096-4d8a-b1e6-f954eb7de59f","Type":"ContainerStarted","Data":"40ff1e9fb5add3ebda7a8f83ccb871c0569d42221df5ba2d0778bbeb09266336"} Oct 11 03:17:46 crc kubenswrapper[4754]: I1011 03:17:46.943981 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49m5p" event={"ID":"70ca9b96-8096-4d8a-b1e6-f954eb7de59f","Type":"ContainerStarted","Data":"3b3db9e461516a0c34347d2322f4b13be3e0b770ece6d07ec832feec1c700ab0"} Oct 11 03:17:46 crc kubenswrapper[4754]: I1011 03:17:46.944020 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-49m5p" Oct 11 03:17:46 crc kubenswrapper[4754]: I1011 03:17:46.966785 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-49m5p" podStartSLOduration=3.966762794 podStartE2EDuration="3.966762794s" podCreationTimestamp="2025-10-11 03:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:17:46.960166028 +0000 UTC m=+714.519110803" watchObservedRunningTime="2025-10-11 03:17:46.966762794 +0000 UTC m=+714.525707579" Oct 11 03:17:51 crc kubenswrapper[4754]: I1011 03:17:51.975061 4754 generic.go:334] "Generic (PLEG): container finished" podID="02c9fc4b-bec8-4468-b120-08f4db18ed07" containerID="ecd50ce7a1b446ee405f831daa623979eea625b52b9a0ca66b97097c9b56cd04" exitCode=0 Oct 11 03:17:51 crc kubenswrapper[4754]: I1011 03:17:51.975144 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerDied","Data":"ecd50ce7a1b446ee405f831daa623979eea625b52b9a0ca66b97097c9b56cd04"} Oct 11 03:17:51 crc kubenswrapper[4754]: I1011 03:17:51.977717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" event={"ID":"01ff4e21-2909-48d9-a3e6-dae8528a1658","Type":"ContainerStarted","Data":"1f9447f374c11c4b6f625215b308f8db3d3c6eda25461296ecb70c37e10d0138"} Oct 11 03:17:51 crc kubenswrapper[4754]: I1011 03:17:51.977898 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:17:52 crc kubenswrapper[4754]: I1011 03:17:52.036289 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" podStartSLOduration=2.407504111 podStartE2EDuration="9.036270223s" podCreationTimestamp="2025-10-11 03:17:43 +0000 UTC" firstStartedPulling="2025-10-11 03:17:45.03986842 +0000 UTC m=+712.598813215" lastFinishedPulling="2025-10-11 03:17:51.668634542 +0000 UTC m=+719.227579327" observedRunningTime="2025-10-11 03:17:52.028777579 +0000 UTC m=+719.587722374" watchObservedRunningTime="2025-10-11 03:17:52.036270223 +0000 UTC m=+719.595215008" Oct 11 03:17:52 crc kubenswrapper[4754]: I1011 03:17:52.988430 4754 generic.go:334] "Generic (PLEG): container finished" podID="02c9fc4b-bec8-4468-b120-08f4db18ed07" containerID="98f6ab7663c06b0cf2e69ab0b920247a66ebe55b63a2b7cbbf07674c2c14fd7d" exitCode=0 Oct 11 03:17:52 crc kubenswrapper[4754]: I1011 03:17:52.988500 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerDied","Data":"98f6ab7663c06b0cf2e69ab0b920247a66ebe55b63a2b7cbbf07674c2c14fd7d"} Oct 11 03:17:53 crc kubenswrapper[4754]: I1011 03:17:53.996869 4754 generic.go:334] "Generic (PLEG): container finished" podID="02c9fc4b-bec8-4468-b120-08f4db18ed07" containerID="d480266ec0c06b98a9c07ba16c2d2f8a43e3aa3c7f1ca478f7846b2adcc8672d" exitCode=0 Oct 11 03:17:53 crc kubenswrapper[4754]: I1011 03:17:53.997305 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerDied","Data":"d480266ec0c06b98a9c07ba16c2d2f8a43e3aa3c7f1ca478f7846b2adcc8672d"} Oct 11 03:17:55 crc kubenswrapper[4754]: I1011 03:17:55.010527 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"e9b4c14d3afa85484ec8408bfde5a868b230810d5b03812b4b1ae85ef3f3b857"} Oct 11 03:17:55 crc kubenswrapper[4754]: I1011 03:17:55.010940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"b4356ba2398384d2403a2219acdb04ef98574f4b4a09b9eabbe5a91711a2b460"} Oct 11 03:17:55 crc kubenswrapper[4754]: I1011 03:17:55.010953 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"471fda468dbd9952344f91e8aab225d4c3168acd23c1200a254184c0f2869f44"} Oct 11 03:17:55 crc kubenswrapper[4754]: I1011 03:17:55.010983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"61a255cd52f09909ea62a69c141d66d540926adbb4101cee00e51f1b07ac562d"} Oct 11 03:17:55 crc kubenswrapper[4754]: I1011 03:17:55.010995 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"c2ffdfbf682dd5e945aed6d834e6273cf3ced2dac3b6f0ed5196f6fff7bb9873"} Oct 11 03:17:56 crc kubenswrapper[4754]: I1011 03:17:56.022639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wjqkr" event={"ID":"02c9fc4b-bec8-4468-b120-08f4db18ed07","Type":"ContainerStarted","Data":"4a0d645d336af85b240a658616ffaf75287838389bc15ad9ac7ff3864bccfd59"} Oct 11 03:17:56 crc kubenswrapper[4754]: I1011 03:17:56.023105 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:56 crc kubenswrapper[4754]: I1011 03:17:56.048521 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wjqkr" podStartSLOduration=6.107758105 podStartE2EDuration="13.048505103s" podCreationTimestamp="2025-10-11 03:17:43 +0000 UTC" firstStartedPulling="2025-10-11 03:17:44.719958148 +0000 UTC m=+712.278902933" lastFinishedPulling="2025-10-11 03:17:51.660705146 +0000 UTC m=+719.219649931" observedRunningTime="2025-10-11 03:17:56.045428609 +0000 UTC m=+723.604373394" watchObservedRunningTime="2025-10-11 03:17:56.048505103 +0000 UTC m=+723.607449888" Oct 11 03:17:59 crc kubenswrapper[4754]: I1011 03:17:59.560464 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:17:59 crc kubenswrapper[4754]: I1011 03:17:59.602572 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:18:04 crc kubenswrapper[4754]: I1011 03:18:04.548528 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6rwvw" Oct 11 03:18:04 crc kubenswrapper[4754]: I1011 03:18:04.565712 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wjqkr" Oct 11 03:18:04 crc kubenswrapper[4754]: I1011 03:18:04.932173 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-9s44v" Oct 11 03:18:05 crc kubenswrapper[4754]: I1011 03:18:05.822243 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-49m5p" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.032174 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.033452 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.036788 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7fnqp" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.037264 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.037716 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.109223 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.166658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qdrg\" (UniqueName: \"kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg\") pod \"openstack-operator-index-fwcjb\" (UID: \"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2\") " pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.268395 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qdrg\" (UniqueName: \"kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg\") pod \"openstack-operator-index-fwcjb\" (UID: \"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2\") " pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.290844 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qdrg\" (UniqueName: \"kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg\") pod \"openstack-operator-index-fwcjb\" (UID: \"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2\") " pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.368189 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:09 crc kubenswrapper[4754]: I1011 03:18:09.824685 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:10 crc kubenswrapper[4754]: I1011 03:18:10.114016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwcjb" event={"ID":"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2","Type":"ContainerStarted","Data":"040445ee330bb2812777c7966fdb0e52e7b23f5f5d65501749b7cc76e90903f0"} Oct 11 03:18:12 crc kubenswrapper[4754]: I1011 03:18:12.393636 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.004301 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-52qtt"] Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.006351 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.013906 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-52qtt"] Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.126732 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dpm\" (UniqueName: \"kubernetes.io/projected/2161836a-986e-4ed4-91ed-9132f0582a85-kube-api-access-29dpm\") pod \"openstack-operator-index-52qtt\" (UID: \"2161836a-986e-4ed4-91ed-9132f0582a85\") " pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.151680 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwcjb" event={"ID":"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2","Type":"ContainerStarted","Data":"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7"} Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.174901 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fwcjb" podStartSLOduration=1.9214803740000002 podStartE2EDuration="4.174877623s" podCreationTimestamp="2025-10-11 03:18:09 +0000 UTC" firstStartedPulling="2025-10-11 03:18:09.840422274 +0000 UTC m=+737.399367069" lastFinishedPulling="2025-10-11 03:18:12.093819523 +0000 UTC m=+739.652764318" observedRunningTime="2025-10-11 03:18:13.167861103 +0000 UTC m=+740.726805898" watchObservedRunningTime="2025-10-11 03:18:13.174877623 +0000 UTC m=+740.733822408" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.229449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dpm\" (UniqueName: \"kubernetes.io/projected/2161836a-986e-4ed4-91ed-9132f0582a85-kube-api-access-29dpm\") pod \"openstack-operator-index-52qtt\" (UID: \"2161836a-986e-4ed4-91ed-9132f0582a85\") " pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.253001 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dpm\" (UniqueName: \"kubernetes.io/projected/2161836a-986e-4ed4-91ed-9132f0582a85-kube-api-access-29dpm\") pod \"openstack-operator-index-52qtt\" (UID: \"2161836a-986e-4ed4-91ed-9132f0582a85\") " pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.329339 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:13 crc kubenswrapper[4754]: I1011 03:18:13.570523 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-52qtt"] Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.165777 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52qtt" event={"ID":"2161836a-986e-4ed4-91ed-9132f0582a85","Type":"ContainerStarted","Data":"82465cbb2ab310e52685e124e908c802136e97132b6cddcbac25806560a0b7f9"} Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.165917 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52qtt" event={"ID":"2161836a-986e-4ed4-91ed-9132f0582a85","Type":"ContainerStarted","Data":"337a87645c3b911a701a79ac95d828e92f77f0dc8f6ba12e775d74b949d231c0"} Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.166373 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fwcjb" podUID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" containerName="registry-server" containerID="cri-o://5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7" gracePeriod=2 Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.622135 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.640886 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-52qtt" podStartSLOduration=2.587026581 podStartE2EDuration="2.640855224s" podCreationTimestamp="2025-10-11 03:18:12 +0000 UTC" firstStartedPulling="2025-10-11 03:18:13.577470315 +0000 UTC m=+741.136415100" lastFinishedPulling="2025-10-11 03:18:13.631298958 +0000 UTC m=+741.190243743" observedRunningTime="2025-10-11 03:18:14.193329942 +0000 UTC m=+741.752274747" watchObservedRunningTime="2025-10-11 03:18:14.640855224 +0000 UTC m=+742.199800019" Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.751849 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qdrg\" (UniqueName: \"kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg\") pod \"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2\" (UID: \"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2\") " Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.762157 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg" (OuterVolumeSpecName: "kube-api-access-9qdrg") pod "8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" (UID: "8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2"). InnerVolumeSpecName "kube-api-access-9qdrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:18:14 crc kubenswrapper[4754]: I1011 03:18:14.854133 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qdrg\" (UniqueName: \"kubernetes.io/projected/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2-kube-api-access-9qdrg\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.174166 4754 generic.go:334] "Generic (PLEG): container finished" podID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" containerID="5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7" exitCode=0 Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.174231 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwcjb" Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.174242 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwcjb" event={"ID":"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2","Type":"ContainerDied","Data":"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7"} Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.174327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwcjb" event={"ID":"8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2","Type":"ContainerDied","Data":"040445ee330bb2812777c7966fdb0e52e7b23f5f5d65501749b7cc76e90903f0"} Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.174361 4754 scope.go:117] "RemoveContainer" containerID="5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7" Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.195198 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.201983 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fwcjb"] Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.209872 4754 scope.go:117] "RemoveContainer" containerID="5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7" Oct 11 03:18:15 crc kubenswrapper[4754]: E1011 03:18:15.210834 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7\": container with ID starting with 5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7 not found: ID does not exist" containerID="5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7" Oct 11 03:18:15 crc kubenswrapper[4754]: I1011 03:18:15.210878 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7"} err="failed to get container status \"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7\": rpc error: code = NotFound desc = could not find container \"5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7\": container with ID starting with 5450a3b086792ebd9a325fbb5ef932038dac65c96e26850dfda9e054d85bc0b7 not found: ID does not exist" Oct 11 03:18:17 crc kubenswrapper[4754]: I1011 03:18:17.094113 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" path="/var/lib/kubelet/pods/8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2/volumes" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.376359 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.376696 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" containerID="cri-o://2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818" gracePeriod=30 Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.466011 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.466276 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerName="route-controller-manager" containerID="cri-o://73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe" gracePeriod=30 Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.857197 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.861910 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.919989 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrl7s\" (UniqueName: \"kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s\") pod \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920050 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config\") pod \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920100 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles\") pod \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920128 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca\") pod \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920196 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert\") pod \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca\") pod \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpx8k\" (UniqueName: \"kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k\") pod \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920367 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config\") pod \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\" (UID: \"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920400 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert\") pod \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\" (UID: \"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8\") " Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.920869 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca" (OuterVolumeSpecName: "client-ca") pod "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" (UID: "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.921049 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" (UID: "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.921062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca" (OuterVolumeSpecName: "client-ca") pod "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" (UID: "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.921337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config" (OuterVolumeSpecName: "config") pod "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" (UID: "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.921427 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config" (OuterVolumeSpecName: "config") pod "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" (UID: "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.926445 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" (UID: "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.926836 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k" (OuterVolumeSpecName: "kube-api-access-cpx8k") pod "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" (UID: "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8"). InnerVolumeSpecName "kube-api-access-cpx8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.926913 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s" (OuterVolumeSpecName: "kube-api-access-lrl7s") pod "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" (UID: "e4e1c4b8-e9f1-479d-805f-4760ce24ceb7"). InnerVolumeSpecName "kube-api-access-lrl7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:18:18 crc kubenswrapper[4754]: I1011 03:18:18.928606 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" (UID: "32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021417 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpx8k\" (UniqueName: \"kubernetes.io/projected/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-kube-api-access-cpx8k\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021446 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021455 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021465 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrl7s\" (UniqueName: \"kubernetes.io/projected/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-kube-api-access-lrl7s\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021474 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021483 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021491 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021499 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.021506 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.206485 4754 generic.go:334] "Generic (PLEG): container finished" podID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerID="73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe" exitCode=0 Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.206531 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.206555 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" event={"ID":"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8","Type":"ContainerDied","Data":"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe"} Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.207039 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52" event={"ID":"32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8","Type":"ContainerDied","Data":"dc96b6d42628a34415be0af5ab66deee4794d892daafce7fa3ed8578e5519d28"} Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.207091 4754 scope.go:117] "RemoveContainer" containerID="73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.208683 4754 generic.go:334] "Generic (PLEG): container finished" podID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerID="2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818" exitCode=0 Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.208717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" event={"ID":"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7","Type":"ContainerDied","Data":"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818"} Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.208743 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" event={"ID":"e4e1c4b8-e9f1-479d-805f-4760ce24ceb7","Type":"ContainerDied","Data":"d0c1a2fe3243ea290af96dea6b91229b00da1474f0c25b5c4ec216e3d3b48b41"} Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.208776 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lwgvj" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.230923 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.236457 4754 scope.go:117] "RemoveContainer" containerID="73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe" Oct 11 03:18:19 crc kubenswrapper[4754]: E1011 03:18:19.237040 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe\": container with ID starting with 73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe not found: ID does not exist" containerID="73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.237107 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe"} err="failed to get container status \"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe\": rpc error: code = NotFound desc = could not find container \"73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe\": container with ID starting with 73f78d8afd37bb7688761305db579751b71924ac69d09bc5c888109fbf76dabe not found: ID does not exist" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.237134 4754 scope.go:117] "RemoveContainer" containerID="2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.237933 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lwgvj"] Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.246248 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.250306 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9cc52"] Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.259487 4754 scope.go:117] "RemoveContainer" containerID="2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818" Oct 11 03:18:19 crc kubenswrapper[4754]: E1011 03:18:19.260264 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818\": container with ID starting with 2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818 not found: ID does not exist" containerID="2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.260436 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818"} err="failed to get container status \"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818\": rpc error: code = NotFound desc = could not find container \"2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818\": container with ID starting with 2eb0c4c7ffe53724b9706c9dd6156059f74045d33df6661fb04ee39fc4c85818 not found: ID does not exist" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.585312 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-649bdfd569-p2npm"] Oct 11 03:18:19 crc kubenswrapper[4754]: E1011 03:18:19.585791 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" containerName="registry-server" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.585822 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" containerName="registry-server" Oct 11 03:18:19 crc kubenswrapper[4754]: E1011 03:18:19.585862 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerName="route-controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.585874 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerName="route-controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: E1011 03:18:19.585899 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.585912 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.586161 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" containerName="route-controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.586200 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b3206c3-50ee-47c9-9ea0-e7af4fae4fd2" containerName="registry-server" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.586222 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" containerName="controller-manager" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.587042 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.589758 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.590044 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.590497 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.590691 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.590816 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.590856 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.594365 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649bdfd569-p2npm"] Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.598721 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.632144 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nw6t\" (UniqueName: \"kubernetes.io/projected/d301ec3f-63be-4c5e-9b69-36dfb72bf812-kube-api-access-9nw6t\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.632843 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-proxy-ca-bundles\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.632998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-client-ca\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.633085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d301ec3f-63be-4c5e-9b69-36dfb72bf812-serving-cert\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.633179 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-config\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.734262 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-client-ca\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.734328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d301ec3f-63be-4c5e-9b69-36dfb72bf812-serving-cert\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.734376 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-config\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.734412 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nw6t\" (UniqueName: \"kubernetes.io/projected/d301ec3f-63be-4c5e-9b69-36dfb72bf812-kube-api-access-9nw6t\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.734440 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-proxy-ca-bundles\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.736404 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-proxy-ca-bundles\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.736882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-client-ca\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.737111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d301ec3f-63be-4c5e-9b69-36dfb72bf812-config\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.741716 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d301ec3f-63be-4c5e-9b69-36dfb72bf812-serving-cert\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.752190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nw6t\" (UniqueName: \"kubernetes.io/projected/d301ec3f-63be-4c5e-9b69-36dfb72bf812-kube-api-access-9nw6t\") pod \"controller-manager-649bdfd569-p2npm\" (UID: \"d301ec3f-63be-4c5e-9b69-36dfb72bf812\") " pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:19 crc kubenswrapper[4754]: I1011 03:18:19.907954 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.326242 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649bdfd569-p2npm"] Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.579513 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9"] Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.580260 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.583270 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.583352 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.583897 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.584273 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.584389 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.584582 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.596018 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9"] Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.753684 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-config\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.753798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e6be81-377e-4c03-a540-1765f1d92185-serving-cert\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.753853 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h2j8\" (UniqueName: \"kubernetes.io/projected/92e6be81-377e-4c03-a540-1765f1d92185-kube-api-access-5h2j8\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.753900 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-client-ca\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.855601 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-config\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.855712 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e6be81-377e-4c03-a540-1765f1d92185-serving-cert\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.855769 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h2j8\" (UniqueName: \"kubernetes.io/projected/92e6be81-377e-4c03-a540-1765f1d92185-kube-api-access-5h2j8\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.855820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-client-ca\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.858337 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-config\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.860239 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e6be81-377e-4c03-a540-1765f1d92185-client-ca\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.867691 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e6be81-377e-4c03-a540-1765f1d92185-serving-cert\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.880912 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h2j8\" (UniqueName: \"kubernetes.io/projected/92e6be81-377e-4c03-a540-1765f1d92185-kube-api-access-5h2j8\") pod \"route-controller-manager-84f665c458-gzlh9\" (UID: \"92e6be81-377e-4c03-a540-1765f1d92185\") " pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:20 crc kubenswrapper[4754]: I1011 03:18:20.943835 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.092007 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8" path="/var/lib/kubelet/pods/32a285a1-54aa-4ba5-8a04-9bfffe6ea0e8/volumes" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.093134 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e1c4b8-e9f1-479d-805f-4760ce24ceb7" path="/var/lib/kubelet/pods/e4e1c4b8-e9f1-479d-805f-4760ce24ceb7/volumes" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.224155 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" event={"ID":"d301ec3f-63be-4c5e-9b69-36dfb72bf812","Type":"ContainerStarted","Data":"583ebeb7a47d492aa1aa538743bdec88d6d1160823fb49cd9c8bb7f7c780f4a6"} Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.224205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" event={"ID":"d301ec3f-63be-4c5e-9b69-36dfb72bf812","Type":"ContainerStarted","Data":"848baf2f5063dd09807d35fbd8fe40f4b17becbfdd28fe78be361e479c5fb4b4"} Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.225414 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.238386 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.241788 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-649bdfd569-p2npm" podStartSLOduration=3.241777317 podStartE2EDuration="3.241777317s" podCreationTimestamp="2025-10-11 03:18:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:18:21.24080164 +0000 UTC m=+748.799746425" watchObservedRunningTime="2025-10-11 03:18:21.241777317 +0000 UTC m=+748.800722102" Oct 11 03:18:21 crc kubenswrapper[4754]: I1011 03:18:21.378180 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9"] Oct 11 03:18:22 crc kubenswrapper[4754]: I1011 03:18:22.233717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" event={"ID":"92e6be81-377e-4c03-a540-1765f1d92185","Type":"ContainerStarted","Data":"58d44158caa8ea9ec9d100be0b2a5f8eec52e85b57cf1dc5d16b89bd5c7ebe98"} Oct 11 03:18:22 crc kubenswrapper[4754]: I1011 03:18:22.234152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" event={"ID":"92e6be81-377e-4c03-a540-1765f1d92185","Type":"ContainerStarted","Data":"9296193bef261a07c8143084b862c196725a6b6dd1a660659703c3fc147729bc"} Oct 11 03:18:22 crc kubenswrapper[4754]: I1011 03:18:22.253101 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" podStartSLOduration=4.25308171 podStartE2EDuration="4.25308171s" podCreationTimestamp="2025-10-11 03:18:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:18:22.250425418 +0000 UTC m=+749.809370213" watchObservedRunningTime="2025-10-11 03:18:22.25308171 +0000 UTC m=+749.812026495" Oct 11 03:18:23 crc kubenswrapper[4754]: I1011 03:18:23.239225 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:23 crc kubenswrapper[4754]: I1011 03:18:23.243640 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84f665c458-gzlh9" Oct 11 03:18:23 crc kubenswrapper[4754]: I1011 03:18:23.329687 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:23 crc kubenswrapper[4754]: I1011 03:18:23.329744 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:23 crc kubenswrapper[4754]: I1011 03:18:23.354260 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:24 crc kubenswrapper[4754]: I1011 03:18:24.326720 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-52qtt" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.658379 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw"] Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.660591 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.663506 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qgbs6" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.674810 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw"] Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.834748 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.834840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.835839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5llh7\" (UniqueName: \"kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.937540 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5llh7\" (UniqueName: \"kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.937656 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.937749 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.938301 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.938366 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:25 crc kubenswrapper[4754]: I1011 03:18:25.961057 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5llh7\" (UniqueName: \"kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7\") pod \"2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:26 crc kubenswrapper[4754]: I1011 03:18:26.020084 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:26 crc kubenswrapper[4754]: I1011 03:18:26.445275 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw"] Oct 11 03:18:26 crc kubenswrapper[4754]: W1011 03:18:26.454977 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3f7cf56_ada2_4541_a04a_bf1d56775252.slice/crio-166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9 WatchSource:0}: Error finding container 166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9: Status 404 returned error can't find the container with id 166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9 Oct 11 03:18:27 crc kubenswrapper[4754]: I1011 03:18:27.213676 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 03:18:27 crc kubenswrapper[4754]: I1011 03:18:27.271428 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerID="95b979a32e813ad2d5c786d8016ccb7ae4281d12bd43a9cc4a31d93f1a34fdc4" exitCode=0 Oct 11 03:18:27 crc kubenswrapper[4754]: I1011 03:18:27.271499 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" event={"ID":"c3f7cf56-ada2-4541-a04a-bf1d56775252","Type":"ContainerDied","Data":"95b979a32e813ad2d5c786d8016ccb7ae4281d12bd43a9cc4a31d93f1a34fdc4"} Oct 11 03:18:27 crc kubenswrapper[4754]: I1011 03:18:27.271538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" event={"ID":"c3f7cf56-ada2-4541-a04a-bf1d56775252","Type":"ContainerStarted","Data":"166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9"} Oct 11 03:18:28 crc kubenswrapper[4754]: I1011 03:18:28.282716 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerID="fa977f115621794add21703114183992e0e05acccc3b5f4a63b51dab6451a6d9" exitCode=0 Oct 11 03:18:28 crc kubenswrapper[4754]: I1011 03:18:28.282806 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" event={"ID":"c3f7cf56-ada2-4541-a04a-bf1d56775252","Type":"ContainerDied","Data":"fa977f115621794add21703114183992e0e05acccc3b5f4a63b51dab6451a6d9"} Oct 11 03:18:29 crc kubenswrapper[4754]: I1011 03:18:29.293853 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerID="cc6892babd9b8c330744a57d1a3f089c237473cb2df55d09b05b593e8e0165a0" exitCode=0 Oct 11 03:18:29 crc kubenswrapper[4754]: I1011 03:18:29.293909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" event={"ID":"c3f7cf56-ada2-4541-a04a-bf1d56775252","Type":"ContainerDied","Data":"cc6892babd9b8c330744a57d1a3f089c237473cb2df55d09b05b593e8e0165a0"} Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.655756 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.736678 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.736747 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.803907 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle\") pod \"c3f7cf56-ada2-4541-a04a-bf1d56775252\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.804107 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util\") pod \"c3f7cf56-ada2-4541-a04a-bf1d56775252\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.804161 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5llh7\" (UniqueName: \"kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7\") pod \"c3f7cf56-ada2-4541-a04a-bf1d56775252\" (UID: \"c3f7cf56-ada2-4541-a04a-bf1d56775252\") " Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.804863 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle" (OuterVolumeSpecName: "bundle") pod "c3f7cf56-ada2-4541-a04a-bf1d56775252" (UID: "c3f7cf56-ada2-4541-a04a-bf1d56775252"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.813570 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7" (OuterVolumeSpecName: "kube-api-access-5llh7") pod "c3f7cf56-ada2-4541-a04a-bf1d56775252" (UID: "c3f7cf56-ada2-4541-a04a-bf1d56775252"). InnerVolumeSpecName "kube-api-access-5llh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.823593 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util" (OuterVolumeSpecName: "util") pod "c3f7cf56-ada2-4541-a04a-bf1d56775252" (UID: "c3f7cf56-ada2-4541-a04a-bf1d56775252"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.906008 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.906071 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c3f7cf56-ada2-4541-a04a-bf1d56775252-util\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:30 crc kubenswrapper[4754]: I1011 03:18:30.906092 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5llh7\" (UniqueName: \"kubernetes.io/projected/c3f7cf56-ada2-4541-a04a-bf1d56775252-kube-api-access-5llh7\") on node \"crc\" DevicePath \"\"" Oct 11 03:18:31 crc kubenswrapper[4754]: I1011 03:18:31.312089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" event={"ID":"c3f7cf56-ada2-4541-a04a-bf1d56775252","Type":"ContainerDied","Data":"166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9"} Oct 11 03:18:31 crc kubenswrapper[4754]: I1011 03:18:31.312773 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="166bd8ce57613ebc119658f186f8de56f8599ee0dea9033b2db7ac302a654de9" Oct 11 03:18:31 crc kubenswrapper[4754]: I1011 03:18:31.312904 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.088080 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8"] Oct 11 03:18:38 crc kubenswrapper[4754]: E1011 03:18:38.089140 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="pull" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.089157 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="pull" Oct 11 03:18:38 crc kubenswrapper[4754]: E1011 03:18:38.089180 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="extract" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.089188 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="extract" Oct 11 03:18:38 crc kubenswrapper[4754]: E1011 03:18:38.089201 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="util" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.089208 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="util" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.089348 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f7cf56-ada2-4541-a04a-bf1d56775252" containerName="extract" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.090241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.094420 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qr4zr" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.112052 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8"] Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.137213 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.138611 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.173826 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.224319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdfnh\" (UniqueName: \"kubernetes.io/projected/3ec784c6-1183-4341-8d58-6cdd11b0b854-kube-api-access-jdfnh\") pod \"openstack-operator-controller-operator-7c6467487b-bgwp8\" (UID: \"3ec784c6-1183-4341-8d58-6cdd11b0b854\") " pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.326485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.326936 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrthr\" (UniqueName: \"kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.327066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.327233 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdfnh\" (UniqueName: \"kubernetes.io/projected/3ec784c6-1183-4341-8d58-6cdd11b0b854-kube-api-access-jdfnh\") pod \"openstack-operator-controller-operator-7c6467487b-bgwp8\" (UID: \"3ec784c6-1183-4341-8d58-6cdd11b0b854\") " pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.354349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdfnh\" (UniqueName: \"kubernetes.io/projected/3ec784c6-1183-4341-8d58-6cdd11b0b854-kube-api-access-jdfnh\") pod \"openstack-operator-controller-operator-7c6467487b-bgwp8\" (UID: \"3ec784c6-1183-4341-8d58-6cdd11b0b854\") " pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.413721 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.429339 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrthr\" (UniqueName: \"kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.430005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.430132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.430880 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.431488 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.451902 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrthr\" (UniqueName: \"kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr\") pod \"certified-operators-frf5g\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.469603 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:38 crc kubenswrapper[4754]: I1011 03:18:38.906822 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8"] Oct 11 03:18:39 crc kubenswrapper[4754]: I1011 03:18:39.043805 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:18:39 crc kubenswrapper[4754]: W1011 03:18:39.055151 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aa93960_9707_4a89_8671_2ffe8b621cf9.slice/crio-43a3e2be0eecc1a98e980b95b6bcde2fd39a7b190ef95b764a6513dc89209d0c WatchSource:0}: Error finding container 43a3e2be0eecc1a98e980b95b6bcde2fd39a7b190ef95b764a6513dc89209d0c: Status 404 returned error can't find the container with id 43a3e2be0eecc1a98e980b95b6bcde2fd39a7b190ef95b764a6513dc89209d0c Oct 11 03:18:39 crc kubenswrapper[4754]: I1011 03:18:39.380072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" event={"ID":"3ec784c6-1183-4341-8d58-6cdd11b0b854","Type":"ContainerStarted","Data":"7936441125dc70937845dbafe784aac2c98ae6df233d3725f090ef288aa041e9"} Oct 11 03:18:39 crc kubenswrapper[4754]: I1011 03:18:39.390386 4754 generic.go:334] "Generic (PLEG): container finished" podID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerID="471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55" exitCode=0 Oct 11 03:18:39 crc kubenswrapper[4754]: I1011 03:18:39.390469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerDied","Data":"471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55"} Oct 11 03:18:39 crc kubenswrapper[4754]: I1011 03:18:39.390518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerStarted","Data":"43a3e2be0eecc1a98e980b95b6bcde2fd39a7b190ef95b764a6513dc89209d0c"} Oct 11 03:18:43 crc kubenswrapper[4754]: I1011 03:18:43.420835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" event={"ID":"3ec784c6-1183-4341-8d58-6cdd11b0b854","Type":"ContainerStarted","Data":"a68203424ee62f93df960854a564d8a9fe65866ba7eb05a908bdbddfa36da670"} Oct 11 03:18:43 crc kubenswrapper[4754]: I1011 03:18:43.423801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerStarted","Data":"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb"} Oct 11 03:18:44 crc kubenswrapper[4754]: I1011 03:18:44.438349 4754 generic.go:334] "Generic (PLEG): container finished" podID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerID="ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb" exitCode=0 Oct 11 03:18:44 crc kubenswrapper[4754]: I1011 03:18:44.438430 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerDied","Data":"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb"} Oct 11 03:18:46 crc kubenswrapper[4754]: I1011 03:18:46.455519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerStarted","Data":"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7"} Oct 11 03:18:46 crc kubenswrapper[4754]: I1011 03:18:46.458255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" event={"ID":"3ec784c6-1183-4341-8d58-6cdd11b0b854","Type":"ContainerStarted","Data":"f308ef4de61d80980fd14fc408f20397cbe93bf7571c2d80bb8551ea84a0f0c7"} Oct 11 03:18:46 crc kubenswrapper[4754]: I1011 03:18:46.458357 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:46 crc kubenswrapper[4754]: I1011 03:18:46.493522 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-frf5g" podStartSLOduration=2.26543789 podStartE2EDuration="8.493496219s" podCreationTimestamp="2025-10-11 03:18:38 +0000 UTC" firstStartedPulling="2025-10-11 03:18:39.4132428 +0000 UTC m=+766.972187585" lastFinishedPulling="2025-10-11 03:18:45.641301119 +0000 UTC m=+773.200245914" observedRunningTime="2025-10-11 03:18:46.486584641 +0000 UTC m=+774.045529446" watchObservedRunningTime="2025-10-11 03:18:46.493496219 +0000 UTC m=+774.052441014" Oct 11 03:18:46 crc kubenswrapper[4754]: I1011 03:18:46.525163 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" podStartSLOduration=1.792010074 podStartE2EDuration="8.525136869s" podCreationTimestamp="2025-10-11 03:18:38 +0000 UTC" firstStartedPulling="2025-10-11 03:18:38.910105866 +0000 UTC m=+766.469050641" lastFinishedPulling="2025-10-11 03:18:45.643232641 +0000 UTC m=+773.202177436" observedRunningTime="2025-10-11 03:18:46.522919078 +0000 UTC m=+774.081863873" watchObservedRunningTime="2025-10-11 03:18:46.525136869 +0000 UTC m=+774.084081664" Oct 11 03:18:48 crc kubenswrapper[4754]: I1011 03:18:48.416853 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7c6467487b-bgwp8" Oct 11 03:18:48 crc kubenswrapper[4754]: I1011 03:18:48.469712 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:48 crc kubenswrapper[4754]: I1011 03:18:48.469806 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:48 crc kubenswrapper[4754]: I1011 03:18:48.523210 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:58 crc kubenswrapper[4754]: I1011 03:18:58.569708 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:18:58 crc kubenswrapper[4754]: I1011 03:18:58.632812 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:18:59 crc kubenswrapper[4754]: I1011 03:18:59.547119 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-frf5g" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="registry-server" containerID="cri-o://9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7" gracePeriod=2 Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.046815 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.170038 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrthr\" (UniqueName: \"kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr\") pod \"4aa93960-9707-4a89-8671-2ffe8b621cf9\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.170218 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content\") pod \"4aa93960-9707-4a89-8671-2ffe8b621cf9\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.170293 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities\") pod \"4aa93960-9707-4a89-8671-2ffe8b621cf9\" (UID: \"4aa93960-9707-4a89-8671-2ffe8b621cf9\") " Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.171734 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities" (OuterVolumeSpecName: "utilities") pod "4aa93960-9707-4a89-8671-2ffe8b621cf9" (UID: "4aa93960-9707-4a89-8671-2ffe8b621cf9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.177061 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr" (OuterVolumeSpecName: "kube-api-access-wrthr") pod "4aa93960-9707-4a89-8671-2ffe8b621cf9" (UID: "4aa93960-9707-4a89-8671-2ffe8b621cf9"). InnerVolumeSpecName "kube-api-access-wrthr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.219626 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4aa93960-9707-4a89-8671-2ffe8b621cf9" (UID: "4aa93960-9707-4a89-8671-2ffe8b621cf9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.272216 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.272249 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa93960-9707-4a89-8671-2ffe8b621cf9-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.272261 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrthr\" (UniqueName: \"kubernetes.io/projected/4aa93960-9707-4a89-8671-2ffe8b621cf9-kube-api-access-wrthr\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.558717 4754 generic.go:334] "Generic (PLEG): container finished" podID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerID="9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7" exitCode=0 Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.558827 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerDied","Data":"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7"} Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.559218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frf5g" event={"ID":"4aa93960-9707-4a89-8671-2ffe8b621cf9","Type":"ContainerDied","Data":"43a3e2be0eecc1a98e980b95b6bcde2fd39a7b190ef95b764a6513dc89209d0c"} Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.558850 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frf5g" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.559287 4754 scope.go:117] "RemoveContainer" containerID="9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.580152 4754 scope.go:117] "RemoveContainer" containerID="ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.596956 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.599814 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-frf5g"] Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.610892 4754 scope.go:117] "RemoveContainer" containerID="471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.627664 4754 scope.go:117] "RemoveContainer" containerID="9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7" Oct 11 03:19:00 crc kubenswrapper[4754]: E1011 03:19:00.628163 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7\": container with ID starting with 9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7 not found: ID does not exist" containerID="9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.628197 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7"} err="failed to get container status \"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7\": rpc error: code = NotFound desc = could not find container \"9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7\": container with ID starting with 9860b297c8c8fb60afdec9821b5b49c753aa612cbf65bcb6f1436038b00672d7 not found: ID does not exist" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.628221 4754 scope.go:117] "RemoveContainer" containerID="ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb" Oct 11 03:19:00 crc kubenswrapper[4754]: E1011 03:19:00.628565 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb\": container with ID starting with ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb not found: ID does not exist" containerID="ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.628611 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb"} err="failed to get container status \"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb\": rpc error: code = NotFound desc = could not find container \"ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb\": container with ID starting with ce574234cf7db9b328c7556416dda90dacd6fc18527e51bd6fb33663c6125cbb not found: ID does not exist" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.628642 4754 scope.go:117] "RemoveContainer" containerID="471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55" Oct 11 03:19:00 crc kubenswrapper[4754]: E1011 03:19:00.628982 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55\": container with ID starting with 471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55 not found: ID does not exist" containerID="471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.629015 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55"} err="failed to get container status \"471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55\": rpc error: code = NotFound desc = could not find container \"471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55\": container with ID starting with 471d1f7c0aa2c4da0d32f078f999043121714f1ffa281b20798029cf76216f55 not found: ID does not exist" Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.736806 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:19:00 crc kubenswrapper[4754]: I1011 03:19:00.736896 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:19:01 crc kubenswrapper[4754]: I1011 03:19:01.092581 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" path="/var/lib/kubelet/pods/4aa93960-9707-4a89-8671-2ffe8b621cf9/volumes" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.075819 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp"] Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.076724 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="registry-server" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.076740 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="registry-server" Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.076758 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="extract-utilities" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.076766 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="extract-utilities" Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.076783 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="extract-content" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.076791 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="extract-content" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.076915 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa93960-9707-4a89-8671-2ffe8b621cf9" containerName="registry-server" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.077643 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.079916 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.080244 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-pp4h9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.081071 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.085232 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8nzqd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.092102 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.093245 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.094839 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4dspl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.102492 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.110376 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.113633 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.151752 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.153051 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.154649 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.155442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.159388 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rggbz" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.162674 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-qjn9m" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.168564 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.195069 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.196665 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.200042 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.205464 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8tblq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.206027 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.219178 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.229899 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.234397 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.235244 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg4jb\" (UniqueName: \"kubernetes.io/projected/64bb612c-78b4-4602-9531-358ffc64b40c-kube-api-access-fg4jb\") pod \"designate-operator-controller-manager-687df44cdb-5fgfl\" (UID: \"64bb612c-78b4-4602-9531-358ffc64b40c\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.235277 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mprd6\" (UniqueName: \"kubernetes.io/projected/aae84d6e-3560-4f7b-85c4-bd3e34e98cae-kube-api-access-mprd6\") pod \"cinder-operator-controller-manager-54874b48b8-8vcz8\" (UID: \"aae84d6e-3560-4f7b-85c4-bd3e34e98cae\") " pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.235328 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlbnz\" (UniqueName: \"kubernetes.io/projected/303c4a96-7b77-49b6-8e02-8a23fefaae79-kube-api-access-wlbnz\") pod \"barbican-operator-controller-manager-64f84fcdbb-6b4pp\" (UID: \"303c4a96-7b77-49b6-8e02-8a23fefaae79\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.237819 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-tgd29" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.238040 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.251332 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.252796 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.254675 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.255101 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rz8rq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.265237 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.266632 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.270500 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-twt9w" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.282212 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.285498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.286069 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.290446 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rwb95" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.311049 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.338971 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342308 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7n6l\" (UniqueName: \"kubernetes.io/projected/c25997fd-2525-4846-a923-38d089a829d7-kube-api-access-t7n6l\") pod \"horizon-operator-controller-manager-6d74794d9b-dk7sg\" (UID: \"c25997fd-2525-4846-a923-38d089a829d7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342397 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg4jb\" (UniqueName: \"kubernetes.io/projected/64bb612c-78b4-4602-9531-358ffc64b40c-kube-api-access-fg4jb\") pod \"designate-operator-controller-manager-687df44cdb-5fgfl\" (UID: \"64bb612c-78b4-4602-9531-358ffc64b40c\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mprd6\" (UniqueName: \"kubernetes.io/projected/aae84d6e-3560-4f7b-85c4-bd3e34e98cae-kube-api-access-mprd6\") pod \"cinder-operator-controller-manager-54874b48b8-8vcz8\" (UID: \"aae84d6e-3560-4f7b-85c4-bd3e34e98cae\") " pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342496 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk75s\" (UniqueName: \"kubernetes.io/projected/9db99c94-e7a3-4762-a719-cd7f92b477a6-kube-api-access-zk75s\") pod \"heat-operator-controller-manager-6d9967f8dd-mbv4c\" (UID: \"9db99c94-e7a3-4762-a719-cd7f92b477a6\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342621 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfbjd\" (UniqueName: \"kubernetes.io/projected/bbeae55c-0565-42d4-877e-2bcc53c30440-kube-api-access-sfbjd\") pod \"glance-operator-controller-manager-7bb46cd7d-4fftb\" (UID: \"bbeae55c-0565-42d4-877e-2bcc53c30440\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342667 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlbnz\" (UniqueName: \"kubernetes.io/projected/303c4a96-7b77-49b6-8e02-8a23fefaae79-kube-api-access-wlbnz\") pod \"barbican-operator-controller-manager-64f84fcdbb-6b4pp\" (UID: \"303c4a96-7b77-49b6-8e02-8a23fefaae79\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.342706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfjm6\" (UniqueName: \"kubernetes.io/projected/d4d92b82-f730-4424-984c-0d1ec138e76c-kube-api-access-dfjm6\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.355675 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.356473 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.365017 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.367354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.383393 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vvc8j" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.390075 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-cc79b" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.399080 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mprd6\" (UniqueName: \"kubernetes.io/projected/aae84d6e-3560-4f7b-85c4-bd3e34e98cae-kube-api-access-mprd6\") pod \"cinder-operator-controller-manager-54874b48b8-8vcz8\" (UID: \"aae84d6e-3560-4f7b-85c4-bd3e34e98cae\") " pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.399489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg4jb\" (UniqueName: \"kubernetes.io/projected/64bb612c-78b4-4602-9531-358ffc64b40c-kube-api-access-fg4jb\") pod \"designate-operator-controller-manager-687df44cdb-5fgfl\" (UID: \"64bb612c-78b4-4602-9531-358ffc64b40c\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.399632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlbnz\" (UniqueName: \"kubernetes.io/projected/303c4a96-7b77-49b6-8e02-8a23fefaae79-kube-api-access-wlbnz\") pod \"barbican-operator-controller-manager-64f84fcdbb-6b4pp\" (UID: \"303c4a96-7b77-49b6-8e02-8a23fefaae79\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.407519 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.425120 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.427082 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7n6l\" (UniqueName: \"kubernetes.io/projected/c25997fd-2525-4846-a923-38d089a829d7-kube-api-access-t7n6l\") pod \"horizon-operator-controller-manager-6d74794d9b-dk7sg\" (UID: \"c25997fd-2525-4846-a923-38d089a829d7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453104 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6rc\" (UniqueName: \"kubernetes.io/projected/5c1d2179-2db4-4e18-9f28-f8fcbd66cdec-kube-api-access-rt6rc\") pod \"manila-operator-controller-manager-59578bc799-h7hgd\" (UID: \"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453140 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453169 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hzl\" (UniqueName: \"kubernetes.io/projected/64a954a8-d5af-48b6-961a-158e5fb739b7-kube-api-access-87hzl\") pod \"keystone-operator-controller-manager-ddb98f99b-dx268\" (UID: \"64a954a8-d5af-48b6-961a-158e5fb739b7\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453196 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49867\" (UniqueName: \"kubernetes.io/projected/ebbf1f59-dc6c-4ebf-99a4-c9f806dce032-kube-api-access-49867\") pod \"ironic-operator-controller-manager-74cb5cbc49-26cn2\" (UID: \"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk75s\" (UniqueName: \"kubernetes.io/projected/9db99c94-e7a3-4762-a719-cd7f92b477a6-kube-api-access-zk75s\") pod \"heat-operator-controller-manager-6d9967f8dd-mbv4c\" (UID: \"9db99c94-e7a3-4762-a719-cd7f92b477a6\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfbjd\" (UniqueName: \"kubernetes.io/projected/bbeae55c-0565-42d4-877e-2bcc53c30440-kube-api-access-sfbjd\") pod \"glance-operator-controller-manager-7bb46cd7d-4fftb\" (UID: \"bbeae55c-0565-42d4-877e-2bcc53c30440\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.453282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfjm6\" (UniqueName: \"kubernetes.io/projected/d4d92b82-f730-4424-984c-0d1ec138e76c-kube-api-access-dfjm6\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.453766 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.453941 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert podName:d4d92b82-f730-4424-984c-0d1ec138e76c nodeName:}" failed. No retries permitted until 2025-10-11 03:19:05.953922473 +0000 UTC m=+793.512867258 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert") pod "infra-operator-controller-manager-585fc5b659-ggjw9" (UID: "d4d92b82-f730-4424-984c-0d1ec138e76c") : secret "infra-operator-webhook-server-cert" not found Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.460036 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.461575 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.473422 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-f8f7h" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.474250 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.475564 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.486742 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-k9gn8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.490937 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk75s\" (UniqueName: \"kubernetes.io/projected/9db99c94-e7a3-4762-a719-cd7f92b477a6-kube-api-access-zk75s\") pod \"heat-operator-controller-manager-6d9967f8dd-mbv4c\" (UID: \"9db99c94-e7a3-4762-a719-cd7f92b477a6\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.501738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7n6l\" (UniqueName: \"kubernetes.io/projected/c25997fd-2525-4846-a923-38d089a829d7-kube-api-access-t7n6l\") pod \"horizon-operator-controller-manager-6d74794d9b-dk7sg\" (UID: \"c25997fd-2525-4846-a923-38d089a829d7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.504607 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfjm6\" (UniqueName: \"kubernetes.io/projected/d4d92b82-f730-4424-984c-0d1ec138e76c-kube-api-access-dfjm6\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.505258 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfbjd\" (UniqueName: \"kubernetes.io/projected/bbeae55c-0565-42d4-877e-2bcc53c30440-kube-api-access-sfbjd\") pod \"glance-operator-controller-manager-7bb46cd7d-4fftb\" (UID: \"bbeae55c-0565-42d4-877e-2bcc53c30440\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.508549 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.527955 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.536899 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.538122 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.541226 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-m876b" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.541505 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.555370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49867\" (UniqueName: \"kubernetes.io/projected/ebbf1f59-dc6c-4ebf-99a4-c9f806dce032-kube-api-access-49867\") pod \"ironic-operator-controller-manager-74cb5cbc49-26cn2\" (UID: \"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.555463 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhbgr\" (UniqueName: \"kubernetes.io/projected/5f0ffe2a-9b82-443b-89d3-be54bdb596b6-kube-api-access-zhbgr\") pod \"neutron-operator-controller-manager-797d478b46-hjcs8\" (UID: \"5f0ffe2a-9b82-443b-89d3-be54bdb596b6\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.555503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsbnf\" (UniqueName: \"kubernetes.io/projected/7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b-kube-api-access-dsbnf\") pod \"mariadb-operator-controller-manager-5777b4f897-t2725\" (UID: \"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.555537 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6rc\" (UniqueName: \"kubernetes.io/projected/5c1d2179-2db4-4e18-9f28-f8fcbd66cdec-kube-api-access-rt6rc\") pod \"manila-operator-controller-manager-59578bc799-h7hgd\" (UID: \"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.555578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hzl\" (UniqueName: \"kubernetes.io/projected/64a954a8-d5af-48b6-961a-158e5fb739b7-kube-api-access-87hzl\") pod \"keystone-operator-controller-manager-ddb98f99b-dx268\" (UID: \"64a954a8-d5af-48b6-961a-158e5fb739b7\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.557750 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.580035 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.588650 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.590604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hzl\" (UniqueName: \"kubernetes.io/projected/64a954a8-d5af-48b6-961a-158e5fb739b7-kube-api-access-87hzl\") pod \"keystone-operator-controller-manager-ddb98f99b-dx268\" (UID: \"64a954a8-d5af-48b6-961a-158e5fb739b7\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.593800 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-78n6h" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.595535 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49867\" (UniqueName: \"kubernetes.io/projected/ebbf1f59-dc6c-4ebf-99a4-c9f806dce032-kube-api-access-49867\") pod \"ironic-operator-controller-manager-74cb5cbc49-26cn2\" (UID: \"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.600328 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6rc\" (UniqueName: \"kubernetes.io/projected/5c1d2179-2db4-4e18-9f28-f8fcbd66cdec-kube-api-access-rt6rc\") pod \"manila-operator-controller-manager-59578bc799-h7hgd\" (UID: \"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.608794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.612576 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.617854 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.646347 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656607 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsbnf\" (UniqueName: \"kubernetes.io/projected/7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b-kube-api-access-dsbnf\") pod \"mariadb-operator-controller-manager-5777b4f897-t2725\" (UID: \"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656721 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94q58\" (UniqueName: \"kubernetes.io/projected/46931d30-913d-4738-86af-3fd45616e694-kube-api-access-94q58\") pod \"octavia-operator-controller-manager-6d7c7ddf95-k92pr\" (UID: \"46931d30-913d-4738-86af-3fd45616e694\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656759 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656781 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwcfb\" (UniqueName: \"kubernetes.io/projected/3654956b-08a4-4add-8706-9a5ab944c7f2-kube-api-access-kwcfb\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhbgr\" (UniqueName: \"kubernetes.io/projected/5f0ffe2a-9b82-443b-89d3-be54bdb596b6-kube-api-access-zhbgr\") pod \"neutron-operator-controller-manager-797d478b46-hjcs8\" (UID: \"5f0ffe2a-9b82-443b-89d3-be54bdb596b6\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.656866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcvlm\" (UniqueName: \"kubernetes.io/projected/1ceda0b1-1af1-4fe9-b522-02d9d9798556-kube-api-access-xcvlm\") pod \"nova-operator-controller-manager-57bb74c7bf-7rvpp\" (UID: \"1ceda0b1-1af1-4fe9-b522-02d9d9798556\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.658943 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.683759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-nl2n4" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.685559 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.686144 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.698308 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.727344 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.732659 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.748826 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-j5hrj" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.749629 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsbnf\" (UniqueName: \"kubernetes.io/projected/7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b-kube-api-access-dsbnf\") pod \"mariadb-operator-controller-manager-5777b4f897-t2725\" (UID: \"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.752364 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.754527 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.756028 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.763777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhbgr\" (UniqueName: \"kubernetes.io/projected/5f0ffe2a-9b82-443b-89d3-be54bdb596b6-kube-api-access-zhbgr\") pod \"neutron-operator-controller-manager-797d478b46-hjcs8\" (UID: \"5f0ffe2a-9b82-443b-89d3-be54bdb596b6\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768139 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768182 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwcfb\" (UniqueName: \"kubernetes.io/projected/3654956b-08a4-4add-8706-9a5ab944c7f2-kube-api-access-kwcfb\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768361 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jtgq\" (UniqueName: \"kubernetes.io/projected/e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e-kube-api-access-9jtgq\") pod \"ovn-operator-controller-manager-869cc7797f-h6mmf\" (UID: \"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768439 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgjmc\" (UniqueName: \"kubernetes.io/projected/8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c-kube-api-access-rgjmc\") pod \"telemetry-operator-controller-manager-578874c84d-kv5hr\" (UID: \"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768465 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcvlm\" (UniqueName: \"kubernetes.io/projected/1ceda0b1-1af1-4fe9-b522-02d9d9798556-kube-api-access-xcvlm\") pod \"nova-operator-controller-manager-57bb74c7bf-7rvpp\" (UID: \"1ceda0b1-1af1-4fe9-b522-02d9d9798556\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncqf6\" (UniqueName: \"kubernetes.io/projected/3bf92f20-fa39-4e8e-8092-f41f691aab74-kube-api-access-ncqf6\") pod \"placement-operator-controller-manager-664664cb68-vxkjq\" (UID: \"3bf92f20-fa39-4e8e-8092-f41f691aab74\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.768635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94q58\" (UniqueName: \"kubernetes.io/projected/46931d30-913d-4738-86af-3fd45616e694-kube-api-access-94q58\") pod \"octavia-operator-controller-manager-6d7c7ddf95-k92pr\" (UID: \"46931d30-913d-4738-86af-3fd45616e694\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.769233 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-j7d8g" Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.769288 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 03:19:05 crc kubenswrapper[4754]: E1011 03:19:05.769346 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert podName:3654956b-08a4-4add-8706-9a5ab944c7f2 nodeName:}" failed. No retries permitted until 2025-10-11 03:19:06.269328745 +0000 UTC m=+793.828273530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" (UID: "3654956b-08a4-4add-8706-9a5ab944c7f2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.775849 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.777583 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.785316 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.786328 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.787505 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.790295 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-jtlgx" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.806727 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.808866 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwcfb\" (UniqueName: \"kubernetes.io/projected/3654956b-08a4-4add-8706-9a5ab944c7f2-kube-api-access-kwcfb\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.838449 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.843802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcvlm\" (UniqueName: \"kubernetes.io/projected/1ceda0b1-1af1-4fe9-b522-02d9d9798556-kube-api-access-xcvlm\") pod \"nova-operator-controller-manager-57bb74c7bf-7rvpp\" (UID: \"1ceda0b1-1af1-4fe9-b522-02d9d9798556\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.865883 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.876951 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94q58\" (UniqueName: \"kubernetes.io/projected/46931d30-913d-4738-86af-3fd45616e694-kube-api-access-94q58\") pod \"octavia-operator-controller-manager-6d7c7ddf95-k92pr\" (UID: \"46931d30-913d-4738-86af-3fd45616e694\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.882754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzcm5\" (UniqueName: \"kubernetes.io/projected/2fe0ff99-f1c3-4446-8d92-653e447b239a-kube-api-access-mzcm5\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nzvxp\" (UID: \"2fe0ff99-f1c3-4446-8d92-653e447b239a\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.882822 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p96l2\" (UniqueName: \"kubernetes.io/projected/8b5b1dbd-7d2f-4716-8a59-354854b70195-kube-api-access-p96l2\") pod \"test-operator-controller-manager-ffcdd6c94-f5rxz\" (UID: \"8b5b1dbd-7d2f-4716-8a59-354854b70195\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.882861 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jtgq\" (UniqueName: \"kubernetes.io/projected/e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e-kube-api-access-9jtgq\") pod \"ovn-operator-controller-manager-869cc7797f-h6mmf\" (UID: \"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.882896 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncqf6\" (UniqueName: \"kubernetes.io/projected/3bf92f20-fa39-4e8e-8092-f41f691aab74-kube-api-access-ncqf6\") pod \"placement-operator-controller-manager-664664cb68-vxkjq\" (UID: \"3bf92f20-fa39-4e8e-8092-f41f691aab74\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.882918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgjmc\" (UniqueName: \"kubernetes.io/projected/8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c-kube-api-access-rgjmc\") pod \"telemetry-operator-controller-manager-578874c84d-kv5hr\" (UID: \"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.884826 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.895350 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.901750 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-qm42g"] Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.903434 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.907916 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.911549 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ms4s8" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.966669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgjmc\" (UniqueName: \"kubernetes.io/projected/8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c-kube-api-access-rgjmc\") pod \"telemetry-operator-controller-manager-578874c84d-kv5hr\" (UID: \"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.967133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncqf6\" (UniqueName: \"kubernetes.io/projected/3bf92f20-fa39-4e8e-8092-f41f691aab74-kube-api-access-ncqf6\") pod \"placement-operator-controller-manager-664664cb68-vxkjq\" (UID: \"3bf92f20-fa39-4e8e-8092-f41f691aab74\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:05 crc kubenswrapper[4754]: I1011 03:19:05.967257 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jtgq\" (UniqueName: \"kubernetes.io/projected/e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e-kube-api-access-9jtgq\") pod \"ovn-operator-controller-manager-869cc7797f-h6mmf\" (UID: \"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.025554 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.029404 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzcm5\" (UniqueName: \"kubernetes.io/projected/2fe0ff99-f1c3-4446-8d92-653e447b239a-kube-api-access-mzcm5\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nzvxp\" (UID: \"2fe0ff99-f1c3-4446-8d92-653e447b239a\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.052299 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p96l2\" (UniqueName: \"kubernetes.io/projected/8b5b1dbd-7d2f-4716-8a59-354854b70195-kube-api-access-p96l2\") pod \"test-operator-controller-manager-ffcdd6c94-f5rxz\" (UID: \"8b5b1dbd-7d2f-4716-8a59-354854b70195\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.052580 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.052676 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrlff\" (UniqueName: \"kubernetes.io/projected/446576e5-02c4-4417-bb50-644f527631dd-kube-api-access-wrlff\") pod \"watcher-operator-controller-manager-646675d848-qm42g\" (UID: \"446576e5-02c4-4417-bb50-644f527631dd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.072545 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-qm42g"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.073421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p96l2\" (UniqueName: \"kubernetes.io/projected/8b5b1dbd-7d2f-4716-8a59-354854b70195-kube-api-access-p96l2\") pod \"test-operator-controller-manager-ffcdd6c94-f5rxz\" (UID: \"8b5b1dbd-7d2f-4716-8a59-354854b70195\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.079516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4d92b82-f730-4424-984c-0d1ec138e76c-cert\") pod \"infra-operator-controller-manager-585fc5b659-ggjw9\" (UID: \"d4d92b82-f730-4424-984c-0d1ec138e76c\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.080521 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.082444 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzcm5\" (UniqueName: \"kubernetes.io/projected/2fe0ff99-f1c3-4446-8d92-653e447b239a-kube-api-access-mzcm5\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nzvxp\" (UID: \"2fe0ff99-f1c3-4446-8d92-653e447b239a\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.140830 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.154178 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrlff\" (UniqueName: \"kubernetes.io/projected/446576e5-02c4-4417-bb50-644f527631dd-kube-api-access-wrlff\") pod \"watcher-operator-controller-manager-646675d848-qm42g\" (UID: \"446576e5-02c4-4417-bb50-644f527631dd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.164905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.180827 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.188718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrlff\" (UniqueName: \"kubernetes.io/projected/446576e5-02c4-4417-bb50-644f527631dd-kube-api-access-wrlff\") pod \"watcher-operator-controller-manager-646675d848-qm42g\" (UID: \"446576e5-02c4-4417-bb50-644f527631dd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.219031 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.220565 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.225211 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.229756 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fs92z" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.230013 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.235471 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.264065 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.264751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4x9k\" (UniqueName: \"kubernetes.io/projected/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-kube-api-access-l4x9k\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.272681 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.273847 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.283644 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wj4zx" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.287627 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.361477 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.366681 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.366735 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnf5v\" (UniqueName: \"kubernetes.io/projected/67767e7e-3d7b-4125-8695-aac89dc68d7f-kube-api-access-nnf5v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm\" (UID: \"67767e7e-3d7b-4125-8695-aac89dc68d7f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.366801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.366835 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4x9k\" (UniqueName: \"kubernetes.io/projected/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-kube-api-access-l4x9k\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: E1011 03:19:06.367331 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 11 03:19:06 crc kubenswrapper[4754]: E1011 03:19:06.367389 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert podName:35cfbe55-6993-4bb0-91d9-a05f8bfad5d4 nodeName:}" failed. No retries permitted until 2025-10-11 03:19:06.867371158 +0000 UTC m=+794.426315943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert") pod "openstack-operator-controller-manager-594bbd78-2lt5n" (UID: "35cfbe55-6993-4bb0-91d9-a05f8bfad5d4") : secret "webhook-server-cert" not found Oct 11 03:19:06 crc kubenswrapper[4754]: E1011 03:19:06.367601 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 03:19:06 crc kubenswrapper[4754]: E1011 03:19:06.367624 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert podName:3654956b-08a4-4add-8706-9a5ab944c7f2 nodeName:}" failed. No retries permitted until 2025-10-11 03:19:07.367617764 +0000 UTC m=+794.926562549 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" (UID: "3654956b-08a4-4add-8706-9a5ab944c7f2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.369893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.373335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.389510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4x9k\" (UniqueName: \"kubernetes.io/projected/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-kube-api-access-l4x9k\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.469751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnf5v\" (UniqueName: \"kubernetes.io/projected/67767e7e-3d7b-4125-8695-aac89dc68d7f-kube-api-access-nnf5v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm\" (UID: \"67767e7e-3d7b-4125-8695-aac89dc68d7f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.503650 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnf5v\" (UniqueName: \"kubernetes.io/projected/67767e7e-3d7b-4125-8695-aac89dc68d7f-kube-api-access-nnf5v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm\" (UID: \"67767e7e-3d7b-4125-8695-aac89dc68d7f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.645897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" event={"ID":"aae84d6e-3560-4f7b-85c4-bd3e34e98cae","Type":"ContainerStarted","Data":"9dcc712453c9131c9abdc1586a0f8df982d7d5b47820698747068f6d96445457"} Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.663883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" event={"ID":"64bb612c-78b4-4602-9531-358ffc64b40c","Type":"ContainerStarted","Data":"86360e6188d4f79f9fdfc59e9bdc60489e547fe8e73eeeffec869fac2df7e033"} Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.735455 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.826572 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.837115 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.875949 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.886513 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35cfbe55-6993-4bb0-91d9-a05f8bfad5d4-cert\") pod \"openstack-operator-controller-manager-594bbd78-2lt5n\" (UID: \"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4\") " pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.943430 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd"] Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.978847 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:06 crc kubenswrapper[4754]: I1011 03:19:06.984532 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.197111 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.206730 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.215451 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.371808 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.390838 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.395732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3654956b-08a4-4add-8706-9a5ab944c7f2-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48\" (UID: \"3654956b-08a4-4add-8706-9a5ab944c7f2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.414006 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.646051 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.669128 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9"] Oct 11 03:19:07 crc kubenswrapper[4754]: W1011 03:19:07.701910 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4d92b82_f730_4424_984c_0d1ec138e76c.slice/crio-11e72824ef8416ed6099d966a71ead7c280671995d2cdff5f4ab19f1db93a14a WatchSource:0}: Error finding container 11e72824ef8416ed6099d966a71ead7c280671995d2cdff5f4ab19f1db93a14a: Status 404 returned error can't find the container with id 11e72824ef8416ed6099d966a71ead7c280671995d2cdff5f4ab19f1db93a14a Oct 11 03:19:07 crc kubenswrapper[4754]: W1011 03:19:07.708122 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2f5b6be_db46_40a9_88f6_5a1fb0d8b30e.slice/crio-7d9feee275fd81f775006781680e45a8fa26b9ed7dae8f360b93b2b90349d0dc WatchSource:0}: Error finding container 7d9feee275fd81f775006781680e45a8fa26b9ed7dae8f360b93b2b90349d0dc: Status 404 returned error can't find the container with id 7d9feee275fd81f775006781680e45a8fa26b9ed7dae8f360b93b2b90349d0dc Oct 11 03:19:07 crc kubenswrapper[4754]: W1011 03:19:07.710268 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e5dfa13_7d80_435e_aa41_a1a6b6a2ca3c.slice/crio-9f2948f53884ea3ab4606ec7f0b247d72adfe767eacd99309a24adb103d19c87 WatchSource:0}: Error finding container 9f2948f53884ea3ab4606ec7f0b247d72adfe767eacd99309a24adb103d19c87: Status 404 returned error can't find the container with id 9f2948f53884ea3ab4606ec7f0b247d72adfe767eacd99309a24adb103d19c87 Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.711634 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" event={"ID":"c25997fd-2525-4846-a923-38d089a829d7","Type":"ContainerStarted","Data":"fff1074eecfca9b740817b5195401bc60b98054c1bcb212b3c988e0a1e39592e"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.711995 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.718680 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" event={"ID":"9db99c94-e7a3-4762-a719-cd7f92b477a6","Type":"ContainerStarted","Data":"124f32f5f47e8553fc14df60242848b69f71170b15e2a60af07776621d1ff944"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.729230 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.730884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" event={"ID":"64a954a8-d5af-48b6-961a-158e5fb739b7","Type":"ContainerStarted","Data":"2b7b87e781b21369109626b46f3f4f448f37ead59d918b370f63dd3895899ac4"} Oct 11 03:19:07 crc kubenswrapper[4754]: W1011 03:19:07.740710 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod303c4a96_7b77_49b6_8e02_8a23fefaae79.slice/crio-9976a3a5719ff46a1373ebcf30dc69433bf011b39a33523fc9cb2a1e8f6fe293 WatchSource:0}: Error finding container 9976a3a5719ff46a1373ebcf30dc69433bf011b39a33523fc9cb2a1e8f6fe293: Status 404 returned error can't find the container with id 9976a3a5719ff46a1373ebcf30dc69433bf011b39a33523fc9cb2a1e8f6fe293 Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.754518 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.764474 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.784596 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.792336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" event={"ID":"3bf92f20-fa39-4e8e-8092-f41f691aab74","Type":"ContainerStarted","Data":"2642b254a2b4719fe6a8509e680b2da2d3447802e611e22ac53f28d2167b5857"} Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.794910 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-49867,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-74cb5cbc49-26cn2_openstack-operators(ebbf1f59-dc6c-4ebf-99a4-c9f806dce032): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.795102 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rgjmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-578874c84d-kv5hr_openstack-operators(8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.795172 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.797573 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" event={"ID":"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b","Type":"ContainerStarted","Data":"5fc2c0ff68ddf5c35c4f3e76e10ddaa206139f87c972e3c44c6dbd0c2e1cc7ef"} Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.798047 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wrlff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-qm42g_openstack-operators(446576e5-02c4-4417-bb50-644f527631dd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.798362 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwcfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48_openstack-operators(3654956b-08a4-4add-8706-9a5ab944c7f2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.801227 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" event={"ID":"5f0ffe2a-9b82-443b-89d3-be54bdb596b6","Type":"ContainerStarted","Data":"482d682303679d91f266de02a4c931078978f7ce6f0ec7f3ce6d766ae6465487"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.803136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" event={"ID":"bbeae55c-0565-42d4-877e-2bcc53c30440","Type":"ContainerStarted","Data":"721b42a7af15bde2a70d82c3bdcd3cb2e6ab1ef0161b884a5fd097ef059406d9"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.807149 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" event={"ID":"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec","Type":"ContainerStarted","Data":"27df069ac7490a7c3e34eb9d9e31010db440dfc32e250f27d8195c0407d261dc"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.811078 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" event={"ID":"46931d30-913d-4738-86af-3fd45616e694","Type":"ContainerStarted","Data":"8e5a389c5feff741bf9bbe2a2000720b2234ce373f19a5362572160333eff511"} Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.812509 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-qm42g"] Oct 11 03:19:07 crc kubenswrapper[4754]: W1011 03:19:07.820199 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b5b1dbd_7d2f_4716_8a59_354854b70195.slice/crio-08e2c5ab5c3979ee52d261f78263068c0a65b5c72e3c36d0e45b48c86c12a5c8 WatchSource:0}: Error finding container 08e2c5ab5c3979ee52d261f78263068c0a65b5c72e3c36d0e45b48c86c12a5c8: Status 404 returned error can't find the container with id 08e2c5ab5c3979ee52d261f78263068c0a65b5c72e3c36d0e45b48c86c12a5c8 Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.821180 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp"] Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.822955 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p96l2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-f5rxz_openstack-operators(8b5b1dbd-7d2f-4716-8a59-354854b70195): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: E1011 03:19:07.824045 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mzcm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-nzvxp_openstack-operators(2fe0ff99-f1c3-4446-8d92-653e447b239a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.835633 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.845733 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n"] Oct 11 03:19:07 crc kubenswrapper[4754]: I1011 03:19:07.851446 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48"] Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.821862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" event={"ID":"1ceda0b1-1af1-4fe9-b522-02d9d9798556","Type":"ContainerStarted","Data":"8c232a598de215fa9b6d3a1d8756cbfc8b1478c8e4c5bf9600e5a5be1208f9cf"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.826841 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" event={"ID":"d4d92b82-f730-4424-984c-0d1ec138e76c","Type":"ContainerStarted","Data":"11e72824ef8416ed6099d966a71ead7c280671995d2cdff5f4ab19f1db93a14a"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.829738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" event={"ID":"303c4a96-7b77-49b6-8e02-8a23fefaae79","Type":"ContainerStarted","Data":"9976a3a5719ff46a1373ebcf30dc69433bf011b39a33523fc9cb2a1e8f6fe293"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.832008 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" event={"ID":"2fe0ff99-f1c3-4446-8d92-653e447b239a","Type":"ContainerStarted","Data":"46d1639f8bb702cd7897409abc72be6efcb0844bddf248a13d6b6ace83380ee8"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.833750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" event={"ID":"8b5b1dbd-7d2f-4716-8a59-354854b70195","Type":"ContainerStarted","Data":"08e2c5ab5c3979ee52d261f78263068c0a65b5c72e3c36d0e45b48c86c12a5c8"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.834888 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" event={"ID":"446576e5-02c4-4417-bb50-644f527631dd","Type":"ContainerStarted","Data":"f34571313a2cf7c1670c34e1dcd726d6c6138fd5f19aac26805f262f1755a072"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.836592 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" event={"ID":"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e","Type":"ContainerStarted","Data":"7d9feee275fd81f775006781680e45a8fa26b9ed7dae8f360b93b2b90349d0dc"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.839056 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" event={"ID":"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4","Type":"ContainerStarted","Data":"a0fdf5e01dfc1354fa8b0d9b7e0f895aeadabb5fa1a6ba85a70bc0eafb7974e4"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.841566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" event={"ID":"3654956b-08a4-4add-8706-9a5ab944c7f2","Type":"ContainerStarted","Data":"4746b7c01491d58d1506844c6b433c28bb41d859952b851f8527211f4ffda4d3"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.843160 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" event={"ID":"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032","Type":"ContainerStarted","Data":"5e0089d19e21b48c832809f85aef41041b7ea95dcdd1bcfa78f5586d67768b67"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.845115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" event={"ID":"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c","Type":"ContainerStarted","Data":"9f2948f53884ea3ab4606ec7f0b247d72adfe767eacd99309a24adb103d19c87"} Oct 11 03:19:08 crc kubenswrapper[4754]: I1011 03:19:08.846836 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" event={"ID":"67767e7e-3d7b-4125-8695-aac89dc68d7f","Type":"ContainerStarted","Data":"ee14d5779cb2f49c707483e842ae26538dd5462d6b6880893d27d88b42b108e3"} Oct 11 03:19:09 crc kubenswrapper[4754]: E1011 03:19:09.334691 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" podUID="8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c" Oct 11 03:19:09 crc kubenswrapper[4754]: E1011 03:19:09.353115 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" podUID="3654956b-08a4-4add-8706-9a5ab944c7f2" Oct 11 03:19:09 crc kubenswrapper[4754]: I1011 03:19:09.855733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" event={"ID":"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c","Type":"ContainerStarted","Data":"9b5327ce54a1a1162f9840fa28763a2fd10ecbabe54454a679c8a29035799291"} Oct 11 03:19:09 crc kubenswrapper[4754]: E1011 03:19:09.857830 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" podUID="8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c" Oct 11 03:19:09 crc kubenswrapper[4754]: I1011 03:19:09.859656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" event={"ID":"3654956b-08a4-4add-8706-9a5ab944c7f2","Type":"ContainerStarted","Data":"a6d947bda1fbcace5b3463a4b6b92d49928a0d8d519b65c6ccb7402a16311b93"} Oct 11 03:19:09 crc kubenswrapper[4754]: E1011 03:19:09.892343 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" podUID="3654956b-08a4-4add-8706-9a5ab944c7f2" Oct 11 03:19:10 crc kubenswrapper[4754]: E1011 03:19:10.870944 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" podUID="8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c" Oct 11 03:19:10 crc kubenswrapper[4754]: E1011 03:19:10.872470 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" podUID="3654956b-08a4-4add-8706-9a5ab944c7f2" Oct 11 03:19:10 crc kubenswrapper[4754]: I1011 03:19:10.993019 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:10.995830 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.005176 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.071844 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.072153 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc5cw\" (UniqueName: \"kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.072208 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.173948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.174044 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc5cw\" (UniqueName: \"kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.174071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.174493 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.174840 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.199630 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc5cw\" (UniqueName: \"kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw\") pod \"redhat-marketplace-pfhzw\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:11 crc kubenswrapper[4754]: I1011 03:19:11.355797 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.168241 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.172341 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.184367 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.228027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.228097 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.228188 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gbt5\" (UniqueName: \"kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.329693 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.329745 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.329809 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gbt5\" (UniqueName: \"kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.330346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.330343 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.352123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gbt5\" (UniqueName: \"kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5\") pod \"community-operators-kxr7n\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:14 crc kubenswrapper[4754]: I1011 03:19:14.497052 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:17 crc kubenswrapper[4754]: E1011 03:19:17.845910 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" podUID="2fe0ff99-f1c3-4446-8d92-653e447b239a" Oct 11 03:19:17 crc kubenswrapper[4754]: I1011 03:19:17.931658 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" event={"ID":"2fe0ff99-f1c3-4446-8d92-653e447b239a","Type":"ContainerStarted","Data":"03e84e355a22ee1e26f2a7299cd6a07723988a88d80e1d994fc56a6e99edfbd0"} Oct 11 03:19:18 crc kubenswrapper[4754]: E1011 03:19:18.057078 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" podUID="446576e5-02c4-4417-bb50-644f527631dd" Oct 11 03:19:18 crc kubenswrapper[4754]: E1011 03:19:18.258292 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" podUID="8b5b1dbd-7d2f-4716-8a59-354854b70195" Oct 11 03:19:18 crc kubenswrapper[4754]: E1011 03:19:18.258391 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" podUID="ebbf1f59-dc6c-4ebf-99a4-c9f806dce032" Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.664609 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.713248 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:18 crc kubenswrapper[4754]: W1011 03:19:18.757668 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75e3ed0c_d1f3_454f_b95b_c06230a2920f.slice/crio-3312ee9f06dffac1b1258bfb0da1280ef9e1f387072e3270d4a0cd36aed41d36 WatchSource:0}: Error finding container 3312ee9f06dffac1b1258bfb0da1280ef9e1f387072e3270d4a0cd36aed41d36: Status 404 returned error can't find the container with id 3312ee9f06dffac1b1258bfb0da1280ef9e1f387072e3270d4a0cd36aed41d36 Oct 11 03:19:18 crc kubenswrapper[4754]: W1011 03:19:18.775591 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff7a8295_2fc0_4547_98eb_0b3880f9b4cc.slice/crio-e0e213ca22ec114604546359c3eeac9b452d721ae8deaf37665ffe6e04ae2392 WatchSource:0}: Error finding container e0e213ca22ec114604546359c3eeac9b452d721ae8deaf37665ffe6e04ae2392: Status 404 returned error can't find the container with id e0e213ca22ec114604546359c3eeac9b452d721ae8deaf37665ffe6e04ae2392 Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.947047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" event={"ID":"67767e7e-3d7b-4125-8695-aac89dc68d7f","Type":"ContainerStarted","Data":"c9917b2c14531412df026edcc7fc1833fa037521485646a89824fdbfd6f98f7b"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.950608 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" event={"ID":"446576e5-02c4-4417-bb50-644f527631dd","Type":"ContainerStarted","Data":"a61e161546843340462d723aa68f08f88cada1f0edad36bd934b6b216af2fe27"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.951947 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" event={"ID":"64a954a8-d5af-48b6-961a-158e5fb739b7","Type":"ContainerStarted","Data":"36cb7fdcbf20b9e70b51e97085696719801abb0b3db7ac3fb48ed8a4bb5b7df7"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.953656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" event={"ID":"64bb612c-78b4-4602-9531-358ffc64b40c","Type":"ContainerStarted","Data":"17de0ffaa2d411cb20dcd47c80ff531fd00fb30023aa2bce4c441020edda813b"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.977669 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" event={"ID":"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4","Type":"ContainerStarted","Data":"6e7c34502384a4e10772341d8726c6f2ff9c889eb32e7d899b343df1b5bcf577"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.977714 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" event={"ID":"35cfbe55-6993-4bb0-91d9-a05f8bfad5d4","Type":"ContainerStarted","Data":"6336b00411062f291327beecd61deb6abe20a4f7f3f98cf28c7c035417ef5bd9"} Oct 11 03:19:18 crc kubenswrapper[4754]: I1011 03:19:18.977762 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.005257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" event={"ID":"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032","Type":"ContainerStarted","Data":"81bc330b9c727acd73028cef1f7dc07d01066fa3947e751923a93edebadb76db"} Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.012811 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm" podStartSLOduration=2.814069531 podStartE2EDuration="13.01278663s" podCreationTimestamp="2025-10-11 03:19:06 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.794483183 +0000 UTC m=+795.353427968" lastFinishedPulling="2025-10-11 03:19:17.993200252 +0000 UTC m=+805.552145067" observedRunningTime="2025-10-11 03:19:18.976507414 +0000 UTC m=+806.535452199" watchObservedRunningTime="2025-10-11 03:19:19.01278663 +0000 UTC m=+806.571731415" Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.027227 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" event={"ID":"8b5b1dbd-7d2f-4716-8a59-354854b70195","Type":"ContainerStarted","Data":"11791a93e881ce76e6f2a7bbabc2d8392945c7bfed6f2b3b1fe97f856825e90a"} Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.031751 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerStarted","Data":"3312ee9f06dffac1b1258bfb0da1280ef9e1f387072e3270d4a0cd36aed41d36"} Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.038149 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerStarted","Data":"e0e213ca22ec114604546359c3eeac9b452d721ae8deaf37665ffe6e04ae2392"} Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.051553 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" event={"ID":"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec","Type":"ContainerStarted","Data":"ac2f5db607ab6ebd00ae96b703acf1604b330243ef89cc081a110e5e1e2472be"} Oct 11 03:19:19 crc kubenswrapper[4754]: I1011 03:19:19.056471 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" podStartSLOduration=13.056446697 podStartE2EDuration="13.056446697s" podCreationTimestamp="2025-10-11 03:19:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:19:19.055708627 +0000 UTC m=+806.614653422" watchObservedRunningTime="2025-10-11 03:19:19.056446697 +0000 UTC m=+806.615391482" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.099165 4754 generic.go:334] "Generic (PLEG): container finished" podID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerID="0cbfacdc28c8fb2f07e31128b676aa45c214c793a0ed4c6a9388cc5514c57d3b" exitCode=0 Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.099736 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerDied","Data":"0cbfacdc28c8fb2f07e31128b676aa45c214c793a0ed4c6a9388cc5514c57d3b"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.117241 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" event={"ID":"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e","Type":"ContainerStarted","Data":"0cef834508669a5807f005231d3a20c2e6fd4bc29257c2fa0572a7c5f66c52ae"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.144066 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerStarted","Data":"c769166f55ae09d20e5801e783e3c9ce6137b4617605e4a20c58fc503198a5ed"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.182855 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" event={"ID":"aae84d6e-3560-4f7b-85c4-bd3e34e98cae","Type":"ContainerStarted","Data":"bb39932ddc537176d4674b15ec0fa665c9fe19d66455ec59d5e28f26a0eb6d3a"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.204159 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" event={"ID":"5c1d2179-2db4-4e18-9f28-f8fcbd66cdec","Type":"ContainerStarted","Data":"f7f1b60bb50edb8730609332a1d96317b0680eaf0a845ed7600220eb1eb0ed56"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.205008 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.238946 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" podStartSLOduration=4.226368146 podStartE2EDuration="15.238926053s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:06.96502362 +0000 UTC m=+794.523968405" lastFinishedPulling="2025-10-11 03:19:17.977581517 +0000 UTC m=+805.536526312" observedRunningTime="2025-10-11 03:19:20.238029229 +0000 UTC m=+807.796974014" watchObservedRunningTime="2025-10-11 03:19:20.238926053 +0000 UTC m=+807.797870838" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.257368 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" event={"ID":"46931d30-913d-4738-86af-3fd45616e694","Type":"ContainerStarted","Data":"705f60d3b08883985aece90251f2ca82802a171fa7a17950ff085d52c0940cbc"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.257439 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" event={"ID":"46931d30-913d-4738-86af-3fd45616e694","Type":"ContainerStarted","Data":"db3ae2f247bc8f8a24f13942ba4e327fad296c7bb007ed10636960c9ae6ca0e7"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.258619 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.275243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" event={"ID":"5f0ffe2a-9b82-443b-89d3-be54bdb596b6","Type":"ContainerStarted","Data":"02f78b9b03d82803da46578a88c3a878ec94d9f577a983b1a542e425ffd5d40e"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.304915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" event={"ID":"1ceda0b1-1af1-4fe9-b522-02d9d9798556","Type":"ContainerStarted","Data":"ef8220ba691a573af5875870b429c0ad5034b781f666eff42f8b6e94cd15cde8"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.309737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" event={"ID":"bbeae55c-0565-42d4-877e-2bcc53c30440","Type":"ContainerStarted","Data":"0bb34c4d18a002e036666aa9407ff71962aed7c4cb7a9fd737ac03ce89683298"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.338142 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" podStartSLOduration=4.4340626610000005 podStartE2EDuration="15.338113029s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.214163391 +0000 UTC m=+794.773108176" lastFinishedPulling="2025-10-11 03:19:18.118213759 +0000 UTC m=+805.677158544" observedRunningTime="2025-10-11 03:19:20.308296378 +0000 UTC m=+807.867241163" watchObservedRunningTime="2025-10-11 03:19:20.338113029 +0000 UTC m=+807.897057814" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.347186 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" event={"ID":"303c4a96-7b77-49b6-8e02-8a23fefaae79","Type":"ContainerStarted","Data":"1447a53c22326c6f667090b0447245bd029bd426a818b93ddd9f1d4c860ebf89"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.370293 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" event={"ID":"64bb612c-78b4-4602-9531-358ffc64b40c","Type":"ContainerStarted","Data":"6d748c835877f2ecb1bebe79abec6921146e5601177443a98a9d56370423c020"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.371498 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.396182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" event={"ID":"d4d92b82-f730-4424-984c-0d1ec138e76c","Type":"ContainerStarted","Data":"3dc175e3384b1aa037143c0aaab47f6db3ad292b0f95762e59928e751a450070"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.397060 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.412329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" event={"ID":"3bf92f20-fa39-4e8e-8092-f41f691aab74","Type":"ContainerStarted","Data":"97f4091f1f6d457353bcebfd224bf0cda63c9078e57e063bbd473e802e232a70"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.426246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" event={"ID":"c25997fd-2525-4846-a923-38d089a829d7","Type":"ContainerStarted","Data":"4e18340c48cee35215b3bb0e3a66d43c664236a04555f92078cc780dd8ab930b"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.463983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" event={"ID":"9db99c94-e7a3-4762-a719-cd7f92b477a6","Type":"ContainerStarted","Data":"675475e6afda80a0d2451965e12b5c057dded95d87e8d24be44908ec5f3f979c"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.497051 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" event={"ID":"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b","Type":"ContainerStarted","Data":"282cede3a1d114c67c6e76038fe19a0e746bb3aae5b1892a334d812026bfc0e0"} Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.499644 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" podStartSLOduration=3.999692715 podStartE2EDuration="15.499632658s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:06.494087571 +0000 UTC m=+794.053032356" lastFinishedPulling="2025-10-11 03:19:17.994027504 +0000 UTC m=+805.552972299" observedRunningTime="2025-10-11 03:19:20.44158623 +0000 UTC m=+808.000531015" watchObservedRunningTime="2025-10-11 03:19:20.499632658 +0000 UTC m=+808.058577433" Oct 11 03:19:20 crc kubenswrapper[4754]: I1011 03:19:20.512913 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" podStartSLOduration=5.035835684 podStartE2EDuration="15.512887298s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.70535355 +0000 UTC m=+795.264298335" lastFinishedPulling="2025-10-11 03:19:18.182405164 +0000 UTC m=+805.741349949" observedRunningTime="2025-10-11 03:19:20.49934838 +0000 UTC m=+808.058293155" watchObservedRunningTime="2025-10-11 03:19:20.512887298 +0000 UTC m=+808.071832083" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.524780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" event={"ID":"1ceda0b1-1af1-4fe9-b522-02d9d9798556","Type":"ContainerStarted","Data":"16d2186278209c233d1a895717cb2ef632c7b612cf40b47ce0aececd63fec81c"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.525555 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.530518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" event={"ID":"303c4a96-7b77-49b6-8e02-8a23fefaae79","Type":"ContainerStarted","Data":"5e75e31477e653fd9f2cf9358daa6682056a1e404442e3a57319b466fc25cda2"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.531327 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.540852 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" podStartSLOduration=6.153099749 podStartE2EDuration="16.540841085s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.793911377 +0000 UTC m=+795.352856162" lastFinishedPulling="2025-10-11 03:19:18.181652713 +0000 UTC m=+805.740597498" observedRunningTime="2025-10-11 03:19:21.539162749 +0000 UTC m=+809.098107554" watchObservedRunningTime="2025-10-11 03:19:21.540841085 +0000 UTC m=+809.099785870" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.542566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerStarted","Data":"187c2051fd304429f84f0e34413296c453e7be17cfdbc4a21e5d92b3463f0886"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.554836 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" event={"ID":"e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e","Type":"ContainerStarted","Data":"7b532ef8341635cbc909dcc46eaedb9e2c8a5ff71b36b71ada853fa363a655a0"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.555007 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.566558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" event={"ID":"5f0ffe2a-9b82-443b-89d3-be54bdb596b6","Type":"ContainerStarted","Data":"c25d90309e702c1e8af8abdcd3431b48765e8be60aee58e80553ae29c998047d"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.567199 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.573651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" event={"ID":"c25997fd-2525-4846-a923-38d089a829d7","Type":"ContainerStarted","Data":"da7b9741a0587db8051513856ec361b7de1383f133d420256864a9a2d0e2e0a3"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.574327 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.580191 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" podStartSLOduration=6.233459343 podStartE2EDuration="16.580175834s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.791947494 +0000 UTC m=+795.350892279" lastFinishedPulling="2025-10-11 03:19:18.138663995 +0000 UTC m=+805.697608770" observedRunningTime="2025-10-11 03:19:21.558576977 +0000 UTC m=+809.117521762" watchObservedRunningTime="2025-10-11 03:19:21.580175834 +0000 UTC m=+809.139120619" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.588210 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" event={"ID":"64a954a8-d5af-48b6-961a-158e5fb739b7","Type":"ContainerStarted","Data":"8c3bbad1dec0e63a00d7feae20a89072d1ad5cf855046dd1b25e43967b442928"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.588461 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.603372 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" event={"ID":"3bf92f20-fa39-4e8e-8092-f41f691aab74","Type":"ContainerStarted","Data":"a269558b91ae81db8f0cf27e0194bd1c8b784f44ac5822dcc7003e3efca2af24"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.603951 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" podStartSLOduration=6.218545948 podStartE2EDuration="16.60393488s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.710992364 +0000 UTC m=+795.269937139" lastFinishedPulling="2025-10-11 03:19:18.096381286 +0000 UTC m=+805.655326071" observedRunningTime="2025-10-11 03:19:21.599107588 +0000 UTC m=+809.158052373" watchObservedRunningTime="2025-10-11 03:19:21.60393488 +0000 UTC m=+809.162879665" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.604664 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.625377 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" event={"ID":"aae84d6e-3560-4f7b-85c4-bd3e34e98cae","Type":"ContainerStarted","Data":"9fabc4a22f9b42ef1070898c9699733c0552ccee2b7719d316f2191ba3571ffc"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.626270 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.628427 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" podStartSLOduration=5.545060834 podStartE2EDuration="16.628407835s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.018890304 +0000 UTC m=+794.577835089" lastFinishedPulling="2025-10-11 03:19:18.102237305 +0000 UTC m=+805.661182090" observedRunningTime="2025-10-11 03:19:21.619276837 +0000 UTC m=+809.178221612" watchObservedRunningTime="2025-10-11 03:19:21.628407835 +0000 UTC m=+809.187352620" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.640442 4754 generic.go:334] "Generic (PLEG): container finished" podID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerID="c769166f55ae09d20e5801e783e3c9ce6137b4617605e4a20c58fc503198a5ed" exitCode=0 Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.640483 4754 generic.go:334] "Generic (PLEG): container finished" podID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerID="048c6b64ceda05eb69ef17d4e7cf2877082281fd27e1ab855dc475c8c0d61753" exitCode=0 Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.640558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerDied","Data":"c769166f55ae09d20e5801e783e3c9ce6137b4617605e4a20c58fc503198a5ed"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.640586 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerDied","Data":"048c6b64ceda05eb69ef17d4e7cf2877082281fd27e1ab855dc475c8c0d61753"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.647187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" event={"ID":"bbeae55c-0565-42d4-877e-2bcc53c30440","Type":"ContainerStarted","Data":"404e86eb1e8dd9882e164083c54096cba20b3847a0721757b653c41a9b486ea6"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.647927 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.648218 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" podStartSLOduration=5.568042368 podStartE2EDuration="16.648185562s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:06.897379032 +0000 UTC m=+794.456323827" lastFinishedPulling="2025-10-11 03:19:17.977522226 +0000 UTC m=+805.536467021" observedRunningTime="2025-10-11 03:19:21.638816088 +0000 UTC m=+809.197760883" watchObservedRunningTime="2025-10-11 03:19:21.648185562 +0000 UTC m=+809.207130357" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.652229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" event={"ID":"d4d92b82-f730-4424-984c-0d1ec138e76c","Type":"ContainerStarted","Data":"821bd1f62886eba7b73842b3b4f8b244f54d709e72d1f33ce98523be64ab44fe"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.660068 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" event={"ID":"9db99c94-e7a3-4762-a719-cd7f92b477a6","Type":"ContainerStarted","Data":"ea5e1117ca10302ac0f809a3e19a336adf1f60efbcc19344dc1a47f65531affb"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.661346 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" podStartSLOduration=5.266811152 podStartE2EDuration="16.661318959s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:06.881792448 +0000 UTC m=+794.440737233" lastFinishedPulling="2025-10-11 03:19:18.276300255 +0000 UTC m=+805.835245040" observedRunningTime="2025-10-11 03:19:21.655256855 +0000 UTC m=+809.214201630" watchObservedRunningTime="2025-10-11 03:19:21.661318959 +0000 UTC m=+809.220263754" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.661857 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.664019 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" event={"ID":"7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b","Type":"ContainerStarted","Data":"8ac9765384067eefa31fc72c22a0f2b68082c767652922bf0842c8946f883ad4"} Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.664757 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.684635 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" podStartSLOduration=5.798772768 podStartE2EDuration="16.684611432s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.216052442 +0000 UTC m=+794.774997227" lastFinishedPulling="2025-10-11 03:19:18.101891106 +0000 UTC m=+805.660835891" observedRunningTime="2025-10-11 03:19:21.682475904 +0000 UTC m=+809.241420689" watchObservedRunningTime="2025-10-11 03:19:21.684611432 +0000 UTC m=+809.243556217" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.713758 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" podStartSLOduration=5.239907802 podStartE2EDuration="16.713728004s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:06.512532173 +0000 UTC m=+794.071476958" lastFinishedPulling="2025-10-11 03:19:17.986352335 +0000 UTC m=+805.545297160" observedRunningTime="2025-10-11 03:19:21.703378802 +0000 UTC m=+809.262323607" watchObservedRunningTime="2025-10-11 03:19:21.713728004 +0000 UTC m=+809.272672789" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.752438 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" podStartSLOduration=6.201266208 podStartE2EDuration="16.752415595s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.647598401 +0000 UTC m=+795.206543186" lastFinishedPulling="2025-10-11 03:19:18.198747788 +0000 UTC m=+805.757692573" observedRunningTime="2025-10-11 03:19:21.743693648 +0000 UTC m=+809.302638433" watchObservedRunningTime="2025-10-11 03:19:21.752415595 +0000 UTC m=+809.311360380" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.783023 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" podStartSLOduration=5.807074174 podStartE2EDuration="16.783003756s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.223219837 +0000 UTC m=+794.782164622" lastFinishedPulling="2025-10-11 03:19:18.199149419 +0000 UTC m=+805.758094204" observedRunningTime="2025-10-11 03:19:21.781721481 +0000 UTC m=+809.340666266" watchObservedRunningTime="2025-10-11 03:19:21.783003756 +0000 UTC m=+809.341948541" Oct 11 03:19:21 crc kubenswrapper[4754]: I1011 03:19:21.787822 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" podStartSLOduration=6.195106671 podStartE2EDuration="16.787811117s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.400444224 +0000 UTC m=+794.959389009" lastFinishedPulling="2025-10-11 03:19:17.99314866 +0000 UTC m=+805.552093455" observedRunningTime="2025-10-11 03:19:21.768024059 +0000 UTC m=+809.326968854" watchObservedRunningTime="2025-10-11 03:19:21.787811117 +0000 UTC m=+809.346755902" Oct 11 03:19:22 crc kubenswrapper[4754]: I1011 03:19:22.678729 4754 generic.go:334] "Generic (PLEG): container finished" podID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerID="187c2051fd304429f84f0e34413296c453e7be17cfdbc4a21e5d92b3463f0886" exitCode=0 Oct 11 03:19:22 crc kubenswrapper[4754]: I1011 03:19:22.678782 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerDied","Data":"187c2051fd304429f84f0e34413296c453e7be17cfdbc4a21e5d92b3463f0886"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.418809 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-54874b48b8-8vcz8" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.430609 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-5fgfl" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.536477 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-dk7sg" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.611887 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-dx268" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.662230 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-h7hgd" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.707212 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-6b4pp" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.716240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" event={"ID":"446576e5-02c4-4417-bb50-644f527631dd","Type":"ContainerStarted","Data":"3f53f75abf3e5de9c2be5e7b7b9ddf215ffdbc9327a7e5874f2ee2949c37a549"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.716328 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.719177 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" event={"ID":"2fe0ff99-f1c3-4446-8d92-653e447b239a","Type":"ContainerStarted","Data":"eb41fcf79111616be70f8bc0d0ac2bc6c0b931bf34d90290c44cf3426e023dad"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.719326 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.721316 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" event={"ID":"3654956b-08a4-4add-8706-9a5ab944c7f2","Type":"ContainerStarted","Data":"4408f560d3275ece970d7b8ab9d7ce50718a91a91bef12eb92a9f8c220b55cdc"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.721536 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.732825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerStarted","Data":"9d02df123cbac1150fed9144a3f52943af9fbc3c5e2fe883ffd09e61162de7d6"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.740615 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" event={"ID":"ebbf1f59-dc6c-4ebf-99a4-c9f806dce032","Type":"ContainerStarted","Data":"57a3432fedd9debeefcc116075d16c92f512e7238c81804005113b711ca82d84"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.741004 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.747136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" event={"ID":"8b5b1dbd-7d2f-4716-8a59-354854b70195","Type":"ContainerStarted","Data":"8b8a008366b7413f3a27173d66820dd30e4c1456a6c947c14990c5811f6c8383"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.747268 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.755651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerStarted","Data":"40ed6bb76e840abc0045df8cdf9761ee9ddb1983acd01316780616e9861600cf"} Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.758336 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" podStartSLOduration=4.548243615 podStartE2EDuration="20.758310313s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.797818143 +0000 UTC m=+795.356762918" lastFinishedPulling="2025-10-11 03:19:24.007884831 +0000 UTC m=+811.566829616" observedRunningTime="2025-10-11 03:19:25.750607353 +0000 UTC m=+813.309552168" watchObservedRunningTime="2025-10-11 03:19:25.758310313 +0000 UTC m=+813.317255098" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.762660 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-t2725" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.789867 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-4fftb" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.792245 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mbv4c" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.792801 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" podStartSLOduration=4.60876414 podStartE2EDuration="20.79278202s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.823929173 +0000 UTC m=+795.382873958" lastFinishedPulling="2025-10-11 03:19:24.007947053 +0000 UTC m=+811.566891838" observedRunningTime="2025-10-11 03:19:25.785091291 +0000 UTC m=+813.344036086" watchObservedRunningTime="2025-10-11 03:19:25.79278202 +0000 UTC m=+813.351726805" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.817845 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" podStartSLOduration=3.978578973 podStartE2EDuration="20.81781857s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.797515465 +0000 UTC m=+795.356460250" lastFinishedPulling="2025-10-11 03:19:24.636755052 +0000 UTC m=+812.195699847" observedRunningTime="2025-10-11 03:19:25.81670789 +0000 UTC m=+813.375652675" watchObservedRunningTime="2025-10-11 03:19:25.81781857 +0000 UTC m=+813.376763355" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.845290 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pfhzw" podStartSLOduration=11.991401791 podStartE2EDuration="15.845265566s" podCreationTimestamp="2025-10-11 03:19:10 +0000 UTC" firstStartedPulling="2025-10-11 03:19:20.169520537 +0000 UTC m=+807.728465322" lastFinishedPulling="2025-10-11 03:19:24.023384312 +0000 UTC m=+811.582329097" observedRunningTime="2025-10-11 03:19:25.836655302 +0000 UTC m=+813.395600087" watchObservedRunningTime="2025-10-11 03:19:25.845265566 +0000 UTC m=+813.404210351" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.866734 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kxr7n" podStartSLOduration=7.977961165 podStartE2EDuration="11.866704789s" podCreationTimestamp="2025-10-11 03:19:14 +0000 UTC" firstStartedPulling="2025-10-11 03:19:20.129662783 +0000 UTC m=+807.688607568" lastFinishedPulling="2025-10-11 03:19:24.018406407 +0000 UTC m=+811.577351192" observedRunningTime="2025-10-11 03:19:25.857157299 +0000 UTC m=+813.416102104" watchObservedRunningTime="2025-10-11 03:19:25.866704789 +0000 UTC m=+813.425649564" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.871018 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-hjcs8" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.904648 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" podStartSLOduration=4.699884736 podStartE2EDuration="20.904618529s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.822791332 +0000 UTC m=+795.381736107" lastFinishedPulling="2025-10-11 03:19:24.027525115 +0000 UTC m=+811.586469900" observedRunningTime="2025-10-11 03:19:25.878415317 +0000 UTC m=+813.437360092" watchObservedRunningTime="2025-10-11 03:19:25.904618529 +0000 UTC m=+813.463563324" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.905165 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-k92pr" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.908349 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" podStartSLOduration=4.677404064 podStartE2EDuration="20.90834069s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.794734059 +0000 UTC m=+795.353678844" lastFinishedPulling="2025-10-11 03:19:24.025670685 +0000 UTC m=+811.584615470" observedRunningTime="2025-10-11 03:19:25.901309339 +0000 UTC m=+813.460254124" watchObservedRunningTime="2025-10-11 03:19:25.90834069 +0000 UTC m=+813.467285475" Oct 11 03:19:25 crc kubenswrapper[4754]: I1011 03:19:25.911609 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7rvpp" Oct 11 03:19:26 crc kubenswrapper[4754]: I1011 03:19:26.030706 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vxkjq" Oct 11 03:19:26 crc kubenswrapper[4754]: I1011 03:19:26.173096 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-ggjw9" Oct 11 03:19:26 crc kubenswrapper[4754]: I1011 03:19:26.240874 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-h6mmf" Oct 11 03:19:26 crc kubenswrapper[4754]: I1011 03:19:26.987373 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-594bbd78-2lt5n" Oct 11 03:19:28 crc kubenswrapper[4754]: I1011 03:19:28.795991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" event={"ID":"8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c","Type":"ContainerStarted","Data":"0c2da9be4f9280bc5f8a1016a4ffa5e19fcf10addbe4ab5ba96492ef4ce6ad62"} Oct 11 03:19:28 crc kubenswrapper[4754]: I1011 03:19:28.798152 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:28 crc kubenswrapper[4754]: I1011 03:19:28.825490 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" podStartSLOduration=3.664692532 podStartE2EDuration="23.825466668s" podCreationTimestamp="2025-10-11 03:19:05 +0000 UTC" firstStartedPulling="2025-10-11 03:19:07.794990786 +0000 UTC m=+795.353935571" lastFinishedPulling="2025-10-11 03:19:27.955764922 +0000 UTC m=+815.514709707" observedRunningTime="2025-10-11 03:19:28.817199544 +0000 UTC m=+816.376144349" watchObservedRunningTime="2025-10-11 03:19:28.825466668 +0000 UTC m=+816.384411453" Oct 11 03:19:30 crc kubenswrapper[4754]: I1011 03:19:30.736423 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:19:30 crc kubenswrapper[4754]: I1011 03:19:30.736492 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:19:30 crc kubenswrapper[4754]: I1011 03:19:30.736551 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:19:30 crc kubenswrapper[4754]: I1011 03:19:30.737309 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:19:30 crc kubenswrapper[4754]: I1011 03:19:30.737392 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac" gracePeriod=600 Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.356693 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.356806 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.411695 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.834949 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac" exitCode=0 Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.835025 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac"} Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.837164 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494"} Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.837299 4754 scope.go:117] "RemoveContainer" containerID="e128cdd4ca88e615e86e9ed49efd77735cdc00384f0cfadebedfb5d820e251ec" Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.907507 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:31 crc kubenswrapper[4754]: I1011 03:19:31.968836 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:33 crc kubenswrapper[4754]: I1011 03:19:33.860795 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pfhzw" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="registry-server" containerID="cri-o://9d02df123cbac1150fed9144a3f52943af9fbc3c5e2fe883ffd09e61162de7d6" gracePeriod=2 Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.069357 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.071379 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.079769 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.147773 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.148051 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.148101 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgnh7\" (UniqueName: \"kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.250858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.251723 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.251032 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.253354 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgnh7\" (UniqueName: \"kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.256815 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.286943 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgnh7\" (UniqueName: \"kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7\") pod \"redhat-operators-jzqrg\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.408855 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.528471 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.529018 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.580680 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.871924 4754 generic.go:334] "Generic (PLEG): container finished" podID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerID="9d02df123cbac1150fed9144a3f52943af9fbc3c5e2fe883ffd09e61162de7d6" exitCode=0 Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.872019 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerDied","Data":"9d02df123cbac1150fed9144a3f52943af9fbc3c5e2fe883ffd09e61162de7d6"} Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.930380 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:34 crc kubenswrapper[4754]: I1011 03:19:34.944539 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:35 crc kubenswrapper[4754]: I1011 03:19:35.882464 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerStarted","Data":"95f7fd24ab9d7750b0fcf10b826f96c3f9c0cf1b957b90ab557b5bb43475ec4a"} Oct 11 03:19:35 crc kubenswrapper[4754]: I1011 03:19:35.891268 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-26cn2" Oct 11 03:19:36 crc kubenswrapper[4754]: I1011 03:19:36.085665 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-kv5hr" Oct 11 03:19:36 crc kubenswrapper[4754]: I1011 03:19:36.146706 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nzvxp" Oct 11 03:19:36 crc kubenswrapper[4754]: I1011 03:19:36.185374 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-f5rxz" Oct 11 03:19:36 crc kubenswrapper[4754]: I1011 03:19:36.376420 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-qm42g" Oct 11 03:19:36 crc kubenswrapper[4754]: I1011 03:19:36.869340 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:37 crc kubenswrapper[4754]: I1011 03:19:37.427451 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48" Oct 11 03:19:37 crc kubenswrapper[4754]: I1011 03:19:37.904326 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kxr7n" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="registry-server" containerID="cri-o://40ed6bb76e840abc0045df8cdf9761ee9ddb1983acd01316780616e9861600cf" gracePeriod=2 Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.253882 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.344069 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc5cw\" (UniqueName: \"kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw\") pod \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.345483 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content\") pod \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.345523 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities\") pod \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\" (UID: \"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc\") " Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.346633 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities" (OuterVolumeSpecName: "utilities") pod "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" (UID: "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.353556 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw" (OuterVolumeSpecName: "kube-api-access-xc5cw") pod "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" (UID: "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc"). InnerVolumeSpecName "kube-api-access-xc5cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.359888 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" (UID: "ff7a8295-2fc0-4547-98eb-0b3880f9b4cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.447390 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc5cw\" (UniqueName: \"kubernetes.io/projected/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-kube-api-access-xc5cw\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.447439 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.447453 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.913881 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerID="d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3" exitCode=0 Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.913923 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerDied","Data":"d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3"} Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.917335 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhzw" event={"ID":"ff7a8295-2fc0-4547-98eb-0b3880f9b4cc","Type":"ContainerDied","Data":"e0e213ca22ec114604546359c3eeac9b452d721ae8deaf37665ffe6e04ae2392"} Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.917387 4754 scope.go:117] "RemoveContainer" containerID="9d02df123cbac1150fed9144a3f52943af9fbc3c5e2fe883ffd09e61162de7d6" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.917412 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhzw" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.938844 4754 scope.go:117] "RemoveContainer" containerID="048c6b64ceda05eb69ef17d4e7cf2877082281fd27e1ab855dc475c8c0d61753" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.963920 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.977200 4754 scope.go:117] "RemoveContainer" containerID="c769166f55ae09d20e5801e783e3c9ce6137b4617605e4a20c58fc503198a5ed" Oct 11 03:19:38 crc kubenswrapper[4754]: I1011 03:19:38.981086 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhzw"] Oct 11 03:19:39 crc kubenswrapper[4754]: I1011 03:19:39.093299 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" path="/var/lib/kubelet/pods/ff7a8295-2fc0-4547-98eb-0b3880f9b4cc/volumes" Oct 11 03:19:39 crc kubenswrapper[4754]: I1011 03:19:39.929032 4754 generic.go:334] "Generic (PLEG): container finished" podID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerID="40ed6bb76e840abc0045df8cdf9761ee9ddb1983acd01316780616e9861600cf" exitCode=0 Oct 11 03:19:39 crc kubenswrapper[4754]: I1011 03:19:39.929087 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerDied","Data":"40ed6bb76e840abc0045df8cdf9761ee9ddb1983acd01316780616e9861600cf"} Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.308949 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.382169 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content\") pod \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.382374 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities\") pod \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.382404 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gbt5\" (UniqueName: \"kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5\") pod \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\" (UID: \"75e3ed0c-d1f3-454f-b95b-c06230a2920f\") " Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.384431 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities" (OuterVolumeSpecName: "utilities") pod "75e3ed0c-d1f3-454f-b95b-c06230a2920f" (UID: "75e3ed0c-d1f3-454f-b95b-c06230a2920f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.389419 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5" (OuterVolumeSpecName: "kube-api-access-5gbt5") pod "75e3ed0c-d1f3-454f-b95b-c06230a2920f" (UID: "75e3ed0c-d1f3-454f-b95b-c06230a2920f"). InnerVolumeSpecName "kube-api-access-5gbt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.431393 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75e3ed0c-d1f3-454f-b95b-c06230a2920f" (UID: "75e3ed0c-d1f3-454f-b95b-c06230a2920f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.483950 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.484003 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gbt5\" (UniqueName: \"kubernetes.io/projected/75e3ed0c-d1f3-454f-b95b-c06230a2920f-kube-api-access-5gbt5\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.484017 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75e3ed0c-d1f3-454f-b95b-c06230a2920f-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.938389 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxr7n" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.938384 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxr7n" event={"ID":"75e3ed0c-d1f3-454f-b95b-c06230a2920f","Type":"ContainerDied","Data":"3312ee9f06dffac1b1258bfb0da1280ef9e1f387072e3270d4a0cd36aed41d36"} Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.938483 4754 scope.go:117] "RemoveContainer" containerID="40ed6bb76e840abc0045df8cdf9761ee9ddb1983acd01316780616e9861600cf" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.940387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerStarted","Data":"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b"} Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.958027 4754 scope.go:117] "RemoveContainer" containerID="187c2051fd304429f84f0e34413296c453e7be17cfdbc4a21e5d92b3463f0886" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.987136 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.988070 4754 scope.go:117] "RemoveContainer" containerID="0cbfacdc28c8fb2f07e31128b676aa45c214c793a0ed4c6a9388cc5514c57d3b" Oct 11 03:19:40 crc kubenswrapper[4754]: I1011 03:19:40.990564 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kxr7n"] Oct 11 03:19:41 crc kubenswrapper[4754]: I1011 03:19:41.094078 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" path="/var/lib/kubelet/pods/75e3ed0c-d1f3-454f-b95b-c06230a2920f/volumes" Oct 11 03:19:41 crc kubenswrapper[4754]: I1011 03:19:41.955628 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerID="a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b" exitCode=0 Oct 11 03:19:41 crc kubenswrapper[4754]: I1011 03:19:41.955725 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerDied","Data":"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b"} Oct 11 03:19:42 crc kubenswrapper[4754]: I1011 03:19:42.965797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerStarted","Data":"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1"} Oct 11 03:19:43 crc kubenswrapper[4754]: I1011 03:19:43.005569 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzqrg" podStartSLOduration=6.6020538890000005 podStartE2EDuration="9.005548012s" podCreationTimestamp="2025-10-11 03:19:34 +0000 UTC" firstStartedPulling="2025-10-11 03:19:39.931644086 +0000 UTC m=+827.490588871" lastFinishedPulling="2025-10-11 03:19:42.335138159 +0000 UTC m=+829.894082994" observedRunningTime="2025-10-11 03:19:43.004479061 +0000 UTC m=+830.563423846" watchObservedRunningTime="2025-10-11 03:19:43.005548012 +0000 UTC m=+830.564492797" Oct 11 03:19:44 crc kubenswrapper[4754]: I1011 03:19:44.410054 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:44 crc kubenswrapper[4754]: I1011 03:19:44.410121 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:45 crc kubenswrapper[4754]: I1011 03:19:45.472655 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzqrg" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="registry-server" probeResult="failure" output=< Oct 11 03:19:45 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 11 03:19:45 crc kubenswrapper[4754]: > Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.388795 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389860 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="extract-utilities" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.389873 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="extract-utilities" Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389883 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.389891 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389907 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="extract-content" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.389914 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="extract-content" Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389927 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="extract-content" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.389933 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="extract-content" Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389949 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="extract-utilities" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.389955 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="extract-utilities" Oct 11 03:19:54 crc kubenswrapper[4754]: E1011 03:19:54.389996 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.390001 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.390132 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7a8295-2fc0-4547-98eb-0b3880f9b4cc" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.390146 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e3ed0c-d1f3-454f-b95b-c06230a2920f" containerName="registry-server" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.390897 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.402028 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.402128 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.403084 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-z2f2h" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.408836 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.412453 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.462788 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.515603 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvjrl\" (UniqueName: \"kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.515833 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.522323 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.540214 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.540357 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.545657 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.549594 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617087 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm67q\" (UniqueName: \"kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617113 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617416 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617672 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvjrl\" (UniqueName: \"kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.617896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.637578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvjrl\" (UniqueName: \"kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl\") pod \"dnsmasq-dns-675f4bcbfc-5m5sg\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.707699 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.718628 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.718730 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.718760 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm67q\" (UniqueName: \"kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.719487 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.719763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.732185 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.743349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm67q\" (UniqueName: \"kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q\") pod \"dnsmasq-dns-78dd6ddcc-kd5n7\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:54 crc kubenswrapper[4754]: I1011 03:19:54.867418 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:19:55 crc kubenswrapper[4754]: I1011 03:19:55.194116 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:19:55 crc kubenswrapper[4754]: W1011 03:19:55.198332 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac3519a1_9e51_4225_a874_cfa333876743.slice/crio-196871f00ca90ae060b25e25a3b10bcadfeb4701f07b9ba9c99cbc7065a1e806 WatchSource:0}: Error finding container 196871f00ca90ae060b25e25a3b10bcadfeb4701f07b9ba9c99cbc7065a1e806: Status 404 returned error can't find the container with id 196871f00ca90ae060b25e25a3b10bcadfeb4701f07b9ba9c99cbc7065a1e806 Oct 11 03:19:55 crc kubenswrapper[4754]: I1011 03:19:55.303508 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:19:55 crc kubenswrapper[4754]: W1011 03:19:55.309402 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod379b49b8_d4fe_4608_88ba_34703cb5d6e4.slice/crio-ea0c6d054a67f1bfb02e9d5277a5e3185968ca4a47d7834174668523d7d60805 WatchSource:0}: Error finding container ea0c6d054a67f1bfb02e9d5277a5e3185968ca4a47d7834174668523d7d60805: Status 404 returned error can't find the container with id ea0c6d054a67f1bfb02e9d5277a5e3185968ca4a47d7834174668523d7d60805 Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.064007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" event={"ID":"379b49b8-d4fe-4608-88ba-34703cb5d6e4","Type":"ContainerStarted","Data":"ea0c6d054a67f1bfb02e9d5277a5e3185968ca4a47d7834174668523d7d60805"} Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.065362 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" event={"ID":"ac3519a1-9e51-4225-a874-cfa333876743","Type":"ContainerStarted","Data":"196871f00ca90ae060b25e25a3b10bcadfeb4701f07b9ba9c99cbc7065a1e806"} Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.065528 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzqrg" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="registry-server" containerID="cri-o://83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1" gracePeriod=2 Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.353052 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.393526 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.394737 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.418496 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.449067 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.449137 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.449176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dtpz\" (UniqueName: \"kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.550060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.550477 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dtpz\" (UniqueName: \"kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.550787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.551618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.552136 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.581250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dtpz\" (UniqueName: \"kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz\") pod \"dnsmasq-dns-5ccc8479f9-vb6x4\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.721542 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.772339 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.796335 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.797542 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.810358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.869766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.869846 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv9w8\" (UniqueName: \"kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.869912 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.873856 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities\") pod \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgnh7\" (UniqueName: \"kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7\") pod \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972261 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content\") pod \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\" (UID: \"0c06649e-c38b-4fce-842e-1de3e7a9e65c\") " Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972448 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972503 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv9w8\" (UniqueName: \"kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.972539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.973464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.974039 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities" (OuterVolumeSpecName: "utilities") pod "0c06649e-c38b-4fce-842e-1de3e7a9e65c" (UID: "0c06649e-c38b-4fce-842e-1de3e7a9e65c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:56 crc kubenswrapper[4754]: I1011 03:19:56.974641 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.007547 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7" (OuterVolumeSpecName: "kube-api-access-zgnh7") pod "0c06649e-c38b-4fce-842e-1de3e7a9e65c" (UID: "0c06649e-c38b-4fce-842e-1de3e7a9e65c"). InnerVolumeSpecName "kube-api-access-zgnh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.039527 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv9w8\" (UniqueName: \"kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8\") pod \"dnsmasq-dns-57d769cc4f-pz82b\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.074273 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgnh7\" (UniqueName: \"kubernetes.io/projected/0c06649e-c38b-4fce-842e-1de3e7a9e65c-kube-api-access-zgnh7\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.074317 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.092734 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerID="83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1" exitCode=0 Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.092833 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzqrg" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.124622 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerDied","Data":"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1"} Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.124675 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzqrg" event={"ID":"0c06649e-c38b-4fce-842e-1de3e7a9e65c","Type":"ContainerDied","Data":"95f7fd24ab9d7750b0fcf10b826f96c3f9c0cf1b957b90ab557b5bb43475ec4a"} Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.124700 4754 scope.go:117] "RemoveContainer" containerID="83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.170638 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c06649e-c38b-4fce-842e-1de3e7a9e65c" (UID: "0c06649e-c38b-4fce-842e-1de3e7a9e65c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.180660 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c06649e-c38b-4fce-842e-1de3e7a9e65c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.195500 4754 scope.go:117] "RemoveContainer" containerID="a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.216724 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.239179 4754 scope.go:117] "RemoveContainer" containerID="d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.302450 4754 scope.go:117] "RemoveContainer" containerID="83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1" Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.303012 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1\": container with ID starting with 83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1 not found: ID does not exist" containerID="83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.303078 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1"} err="failed to get container status \"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1\": rpc error: code = NotFound desc = could not find container \"83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1\": container with ID starting with 83e2f63ba4016a6e96cc6f2c3228b4e755b0d30a76a7a598ad2bd2998fd521e1 not found: ID does not exist" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.303102 4754 scope.go:117] "RemoveContainer" containerID="a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b" Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.304073 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b\": container with ID starting with a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b not found: ID does not exist" containerID="a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.304097 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b"} err="failed to get container status \"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b\": rpc error: code = NotFound desc = could not find container \"a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b\": container with ID starting with a7055a3f6cf09de7d76787103f4dbab3a5fe1ca37362601c8efa11756362930b not found: ID does not exist" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.304111 4754 scope.go:117] "RemoveContainer" containerID="d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3" Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.304430 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3\": container with ID starting with d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3 not found: ID does not exist" containerID="d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.304472 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3"} err="failed to get container status \"d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3\": rpc error: code = NotFound desc = could not find container \"d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3\": container with ID starting with d8c4868d5ce803c03071b0e476cf283e12dc04f914b6758f1c708394f46061f3 not found: ID does not exist" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.431103 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.437500 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.442232 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzqrg"] Oct 11 03:19:57 crc kubenswrapper[4754]: W1011 03:19:57.488995 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f2442f3_c008_434a_92ac_1b366deba167.slice/crio-2a6b926ee1ee03ebc5cd3a66ecab8f9c7f4109cfd1a367e901d17bb6dd18d236 WatchSource:0}: Error finding container 2a6b926ee1ee03ebc5cd3a66ecab8f9c7f4109cfd1a367e901d17bb6dd18d236: Status 404 returned error can't find the container with id 2a6b926ee1ee03ebc5cd3a66ecab8f9c7f4109cfd1a367e901d17bb6dd18d236 Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.543099 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.543439 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="registry-server" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.543458 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="registry-server" Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.543485 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="extract-utilities" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.543493 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="extract-utilities" Oct 11 03:19:57 crc kubenswrapper[4754]: E1011 03:19:57.543517 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="extract-content" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.543524 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="extract-content" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.543724 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" containerName="registry-server" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.544632 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.547630 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.547725 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q5psc" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.547763 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.552802 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.553015 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.555761 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.558919 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.559388 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690468 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690545 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690616 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690666 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690693 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b98z8\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690762 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690810 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690838 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.690879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.746129 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:19:57 crc kubenswrapper[4754]: W1011 03:19:57.752326 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod907bf885_81ef_4141_b3f6_77da00f659f4.slice/crio-5de3d6f8e9859a86212396d49e8d6064d7bca522f5bff19f383f87ffb1edae21 WatchSource:0}: Error finding container 5de3d6f8e9859a86212396d49e8d6064d7bca522f5bff19f383f87ffb1edae21: Status 404 returned error can't find the container with id 5de3d6f8e9859a86212396d49e8d6064d7bca522f5bff19f383f87ffb1edae21 Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792579 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792677 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b98z8\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792717 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792735 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792782 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792832 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.792920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.793381 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.793533 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.795332 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.795715 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.796047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.796814 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.799276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.808781 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.808804 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.812466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.813281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b98z8\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.815989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.877611 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.943029 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.944533 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.949986 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.951385 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9nkh9" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.951499 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.951736 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.951804 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.951998 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.952370 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.959855 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.995823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.995903 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.995927 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.995956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.996018 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.996049 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d89xv\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.997275 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.997373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.997450 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.997492 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:57 crc kubenswrapper[4754]: I1011 03:19:57.997647 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.099748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.100398 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.100593 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.101137 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.101149 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103078 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103356 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d89xv\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103510 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103530 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103625 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.103651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.108406 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.111627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.114472 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.118910 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.120565 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.127628 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" event={"ID":"0f2442f3-c008-434a-92ac-1b366deba167","Type":"ContainerStarted","Data":"2a6b926ee1ee03ebc5cd3a66ecab8f9c7f4109cfd1a367e901d17bb6dd18d236"} Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.129785 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" event={"ID":"907bf885-81ef-4141-b3f6-77da00f659f4","Type":"ContainerStarted","Data":"5de3d6f8e9859a86212396d49e8d6064d7bca522f5bff19f383f87ffb1edae21"} Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.133449 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.137951 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d89xv\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.148621 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.172311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.275914 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.478443 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:19:58 crc kubenswrapper[4754]: I1011 03:19:58.786547 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:19:58 crc kubenswrapper[4754]: W1011 03:19:58.807406 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda16af4b4_0dce_4af9_b87e_c9c80560d631.slice/crio-30e8c792249feafbf5309f5a27f0723aec06e4585008c7a8b7d8b001788390f6 WatchSource:0}: Error finding container 30e8c792249feafbf5309f5a27f0723aec06e4585008c7a8b7d8b001788390f6: Status 404 returned error can't find the container with id 30e8c792249feafbf5309f5a27f0723aec06e4585008c7a8b7d8b001788390f6 Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.099047 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c06649e-c38b-4fce-842e-1de3e7a9e65c" path="/var/lib/kubelet/pods/0c06649e-c38b-4fce-842e-1de3e7a9e65c/volumes" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.147905 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerStarted","Data":"4f2ca661286ccbe569ca591e36556b1d6e6ac265cce5b6381c6146f205575346"} Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.155113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerStarted","Data":"30e8c792249feafbf5309f5a27f0723aec06e4585008c7a8b7d8b001788390f6"} Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.198993 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.200180 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.202576 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.221173 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.221196 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.221596 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.221700 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-dpsff" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.281906 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.329403 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.363876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.363934 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kolla-config\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.363982 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-secrets\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364010 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364276 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364346 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xmmf\" (UniqueName: \"kubernetes.io/projected/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kube-api-access-8xmmf\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-default\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.364407 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xmmf\" (UniqueName: \"kubernetes.io/projected/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kube-api-access-8xmmf\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-default\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465420 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465445 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kolla-config\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-secrets\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465518 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.465571 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.470848 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kolla-config\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.471903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.472157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.472420 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.473038 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-config-data-default\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.488813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.497047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.503520 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xmmf\" (UniqueName: \"kubernetes.io/projected/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-kube-api-access-8xmmf\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.508515 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6a40ecec-4fe2-41e1-8d6c-43e7689eedfe-secrets\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.530234 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe\") " pod="openstack/openstack-galera-0" Oct 11 03:19:59 crc kubenswrapper[4754]: I1011 03:19:59.616615 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.372240 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 11 03:20:00 crc kubenswrapper[4754]: W1011 03:20:00.386987 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a40ecec_4fe2_41e1_8d6c_43e7689eedfe.slice/crio-6aa84d5c9d78582a8afa9c8d6aaab21d4c3b15cb549c2309238ffa306141e3c3 WatchSource:0}: Error finding container 6aa84d5c9d78582a8afa9c8d6aaab21d4c3b15cb549c2309238ffa306141e3c3: Status 404 returned error can't find the container with id 6aa84d5c9d78582a8afa9c8d6aaab21d4c3b15cb549c2309238ffa306141e3c3 Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.654973 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.656409 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.659605 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.661050 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.661293 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hh9nd" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.661439 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.664621 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.707883 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.709088 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.712396 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.712669 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-t5hhw" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.712788 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.719887 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.792905 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.792949 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793015 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793036 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793075 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793098 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xssgj\" (UniqueName: \"kubernetes.io/projected/0dc66be1-e62f-4919-8124-9b571ac1849b-kube-api-access-xssgj\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793113 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-config-data\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793137 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793154 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793181 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct45r\" (UniqueName: \"kubernetes.io/projected/01de3ace-e995-419f-ad27-75ee2eb0e110-kube-api-access-ct45r\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793264 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-kolla-config\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.793286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.894283 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.895903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.895975 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896015 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct45r\" (UniqueName: \"kubernetes.io/projected/01de3ace-e995-419f-ad27-75ee2eb0e110-kube-api-access-ct45r\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896061 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896087 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-kolla-config\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896119 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896274 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xssgj\" (UniqueName: \"kubernetes.io/projected/0dc66be1-e62f-4919-8124-9b571ac1849b-kube-api-access-xssgj\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.896381 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-config-data\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.897439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-config-data\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.895829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.898037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.901748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.902085 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01de3ace-e995-419f-ad27-75ee2eb0e110-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.903443 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0dc66be1-e62f-4919-8124-9b571ac1849b-kolla-config\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.905227 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.918166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.918487 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc66be1-e62f-4919-8124-9b571ac1849b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.922890 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.923225 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.923814 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/01de3ace-e995-419f-ad27-75ee2eb0e110-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.926446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xssgj\" (UniqueName: \"kubernetes.io/projected/0dc66be1-e62f-4919-8124-9b571ac1849b-kube-api-access-xssgj\") pod \"memcached-0\" (UID: \"0dc66be1-e62f-4919-8124-9b571ac1849b\") " pod="openstack/memcached-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.930135 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct45r\" (UniqueName: \"kubernetes.io/projected/01de3ace-e995-419f-ad27-75ee2eb0e110-kube-api-access-ct45r\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.960713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01de3ace-e995-419f-ad27-75ee2eb0e110\") " pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:00 crc kubenswrapper[4754]: I1011 03:20:00.996778 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:01 crc kubenswrapper[4754]: I1011 03:20:01.044898 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 11 03:20:01 crc kubenswrapper[4754]: I1011 03:20:01.188147 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe","Type":"ContainerStarted","Data":"6aa84d5c9d78582a8afa9c8d6aaab21d4c3b15cb549c2309238ffa306141e3c3"} Oct 11 03:20:01 crc kubenswrapper[4754]: I1011 03:20:01.621779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 03:20:01 crc kubenswrapper[4754]: W1011 03:20:01.639698 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01de3ace_e995_419f_ad27_75ee2eb0e110.slice/crio-aeaa4407d8548b9b8a972940bce811f6105b51c61a9c6f25f1e77a21e36ae7f5 WatchSource:0}: Error finding container aeaa4407d8548b9b8a972940bce811f6105b51c61a9c6f25f1e77a21e36ae7f5: Status 404 returned error can't find the container with id aeaa4407d8548b9b8a972940bce811f6105b51c61a9c6f25f1e77a21e36ae7f5 Oct 11 03:20:01 crc kubenswrapper[4754]: I1011 03:20:01.850024 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 11 03:20:01 crc kubenswrapper[4754]: W1011 03:20:01.860107 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dc66be1_e62f_4919_8124_9b571ac1849b.slice/crio-3675206d79337202a09895bb21749df62dad1a30381e7d718ffc7b4c77bf3478 WatchSource:0}: Error finding container 3675206d79337202a09895bb21749df62dad1a30381e7d718ffc7b4c77bf3478: Status 404 returned error can't find the container with id 3675206d79337202a09895bb21749df62dad1a30381e7d718ffc7b4c77bf3478 Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.209651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0dc66be1-e62f-4919-8124-9b571ac1849b","Type":"ContainerStarted","Data":"3675206d79337202a09895bb21749df62dad1a30381e7d718ffc7b4c77bf3478"} Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.231504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01de3ace-e995-419f-ad27-75ee2eb0e110","Type":"ContainerStarted","Data":"aeaa4407d8548b9b8a972940bce811f6105b51c61a9c6f25f1e77a21e36ae7f5"} Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.474251 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.475299 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.481644 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cttjr" Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.497586 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.629363 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wxml\" (UniqueName: \"kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml\") pod \"kube-state-metrics-0\" (UID: \"976f8dd9-0b51-4b45-9192-9ddf098acd48\") " pod="openstack/kube-state-metrics-0" Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.733786 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wxml\" (UniqueName: \"kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml\") pod \"kube-state-metrics-0\" (UID: \"976f8dd9-0b51-4b45-9192-9ddf098acd48\") " pod="openstack/kube-state-metrics-0" Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.755064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wxml\" (UniqueName: \"kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml\") pod \"kube-state-metrics-0\" (UID: \"976f8dd9-0b51-4b45-9192-9ddf098acd48\") " pod="openstack/kube-state-metrics-0" Oct 11 03:20:02 crc kubenswrapper[4754]: I1011 03:20:02.827184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.890337 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7scx8"] Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.893480 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.906534 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fg5pv" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.907009 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.907175 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.912394 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8"] Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.918594 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7sm4b"] Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.922586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:06 crc kubenswrapper[4754]: I1011 03:20:06.937883 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7sm4b"] Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.028378 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-etc-ovs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.028448 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/953544e1-701d-4d5c-831e-77e952fc8504-scripts\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.028478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1842e17e-2c29-46ec-bdec-a6802d656e10-scripts\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.028687 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-run\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.028953 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-log-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-lib\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029206 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpm7\" (UniqueName: \"kubernetes.io/projected/1842e17e-2c29-46ec-bdec-a6802d656e10-kube-api-access-jzpm7\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029305 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029368 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-ovn-controller-tls-certs\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029393 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-combined-ca-bundle\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029421 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6tbs\" (UniqueName: \"kubernetes.io/projected/953544e1-701d-4d5c-831e-77e952fc8504-kube-api-access-d6tbs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029444 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.029497 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-log\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130690 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpm7\" (UniqueName: \"kubernetes.io/projected/1842e17e-2c29-46ec-bdec-a6802d656e10-kube-api-access-jzpm7\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-ovn-controller-tls-certs\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-combined-ca-bundle\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130832 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6tbs\" (UniqueName: \"kubernetes.io/projected/953544e1-701d-4d5c-831e-77e952fc8504-kube-api-access-d6tbs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.130888 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-log\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-etc-ovs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131038 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/953544e1-701d-4d5c-831e-77e952fc8504-scripts\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1842e17e-2c29-46ec-bdec-a6802d656e10-scripts\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131122 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-run\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-log-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.131248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-lib\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.134399 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/953544e1-701d-4d5c-831e-77e952fc8504-scripts\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.134822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1842e17e-2c29-46ec-bdec-a6802d656e10-scripts\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135000 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-lib\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135068 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-etc-ovs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-log\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-log-ovn\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135452 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1842e17e-2c29-46ec-bdec-a6802d656e10-var-run\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.135442 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/953544e1-701d-4d5c-831e-77e952fc8504-var-run\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.145466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-combined-ca-bundle\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.147649 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6tbs\" (UniqueName: \"kubernetes.io/projected/953544e1-701d-4d5c-831e-77e952fc8504-kube-api-access-d6tbs\") pod \"ovn-controller-ovs-7sm4b\" (UID: \"953544e1-701d-4d5c-831e-77e952fc8504\") " pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.149435 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1842e17e-2c29-46ec-bdec-a6802d656e10-ovn-controller-tls-certs\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.150015 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpm7\" (UniqueName: \"kubernetes.io/projected/1842e17e-2c29-46ec-bdec-a6802d656e10-kube-api-access-jzpm7\") pod \"ovn-controller-7scx8\" (UID: \"1842e17e-2c29-46ec-bdec-a6802d656e10\") " pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.257355 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.282436 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.731781 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.737817 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.740509 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-s8zb6" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.740594 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.740712 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.741004 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.741481 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.743155 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849119 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849200 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849244 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849292 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzlcp\" (UniqueName: \"kubernetes.io/projected/daa384d9-ff26-4962-872c-68d7ffea94fe-kube-api-access-bzlcp\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849342 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.849360 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950474 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950522 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950701 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzlcp\" (UniqueName: \"kubernetes.io/projected/daa384d9-ff26-4962-872c-68d7ffea94fe-kube-api-access-bzlcp\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.950753 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.951526 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.951815 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.952312 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.952915 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daa384d9-ff26-4962-872c-68d7ffea94fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.958762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.959793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.961672 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa384d9-ff26-4962-872c-68d7ffea94fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.968637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzlcp\" (UniqueName: \"kubernetes.io/projected/daa384d9-ff26-4962-872c-68d7ffea94fe-kube-api-access-bzlcp\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:07 crc kubenswrapper[4754]: I1011 03:20:07.976604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"daa384d9-ff26-4962-872c-68d7ffea94fe\") " pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:08 crc kubenswrapper[4754]: I1011 03:20:08.059606 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.824382 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.827617 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.833225 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-pm6ks" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.833236 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.833221 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.833844 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.851016 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.885783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.885851 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886164 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zk5z\" (UniqueName: \"kubernetes.io/projected/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-kube-api-access-8zk5z\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886276 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886297 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.886321 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.987872 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.987936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.987989 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988055 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zk5z\" (UniqueName: \"kubernetes.io/projected/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-kube-api-access-8zk5z\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988115 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988137 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.988826 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.989177 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.989404 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.991889 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.998297 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.998604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:09 crc kubenswrapper[4754]: I1011 03:20:09.999088 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:10 crc kubenswrapper[4754]: I1011 03:20:10.013774 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zk5z\" (UniqueName: \"kubernetes.io/projected/5fa3e4c1-2c10-4fd7-950f-380c41b8af4b-kube-api-access-8zk5z\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:10 crc kubenswrapper[4754]: I1011 03:20:10.014732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:10 crc kubenswrapper[4754]: I1011 03:20:10.152639 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.039899 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.040746 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d89xv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(a16af4b4-0dce-4af9-b87e-c9c80560d631): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.042001 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.050667 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.050846 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b98z8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(ea4e5c84-3c0a-40fb-bccd-529f48f8c908): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.052023 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.390378 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" Oct 11 03:20:18 crc kubenswrapper[4754]: E1011 03:20:18.392997 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" Oct 11 03:20:22 crc kubenswrapper[4754]: E1011 03:20:22.171429 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 11 03:20:22 crc kubenswrapper[4754]: E1011 03:20:22.171936 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8xmmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(6a40ecec-4fe2-41e1-8d6c-43e7689eedfe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:22 crc kubenswrapper[4754]: E1011 03:20:22.174140 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="6a40ecec-4fe2-41e1-8d6c-43e7689eedfe" Oct 11 03:20:22 crc kubenswrapper[4754]: E1011 03:20:22.433590 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="6a40ecec-4fe2-41e1-8d6c-43e7689eedfe" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.645238 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.645798 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sm67q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-kd5n7_openstack(379b49b8-d4fe-4608-88ba-34703cb5d6e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.646900 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" podUID="379b49b8-d4fe-4608-88ba-34703cb5d6e4" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.656972 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.657173 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nvjrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-5m5sg_openstack(ac3519a1-9e51-4225-a874-cfa333876743): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.658469 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" podUID="ac3519a1-9e51-4225-a874-cfa333876743" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.663890 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.664006 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8dtpz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-vb6x4_openstack(0f2442f3-c008-434a-92ac-1b366deba167): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.665103 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" podUID="0f2442f3-c008-434a-92ac-1b366deba167" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.673617 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.673705 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fv9w8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-pz82b_openstack(907bf885-81ef-4141-b3f6-77da00f659f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:20:23 crc kubenswrapper[4754]: E1011 03:20:23.674924 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" podUID="907bf885-81ef-4141-b3f6-77da00f659f4" Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.088983 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8"] Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.096706 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.167478 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.228638 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7sm4b"] Oct 11 03:20:24 crc kubenswrapper[4754]: W1011 03:20:24.236116 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod953544e1_701d_4d5c_831e_77e952fc8504.slice/crio-fa7f51de27bf5596a185cc74c2d52ff6ff53d935399b20c30f1c69b31cce8602 WatchSource:0}: Error finding container fa7f51de27bf5596a185cc74c2d52ff6ff53d935399b20c30f1c69b31cce8602: Status 404 returned error can't find the container with id fa7f51de27bf5596a185cc74c2d52ff6ff53d935399b20c30f1c69b31cce8602 Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.334298 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 03:20:24 crc kubenswrapper[4754]: W1011 03:20:24.338644 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fa3e4c1_2c10_4fd7_950f_380c41b8af4b.slice/crio-7c2df06232ce263d629186e038d6f7c98967f1ad1d849ba26faf9859df36e33f WatchSource:0}: Error finding container 7c2df06232ce263d629186e038d6f7c98967f1ad1d849ba26faf9859df36e33f: Status 404 returned error can't find the container with id 7c2df06232ce263d629186e038d6f7c98967f1ad1d849ba26faf9859df36e33f Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.447652 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b","Type":"ContainerStarted","Data":"7c2df06232ce263d629186e038d6f7c98967f1ad1d849ba26faf9859df36e33f"} Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.450254 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0dc66be1-e62f-4919-8124-9b571ac1849b","Type":"ContainerStarted","Data":"28c03066780aaaad780f948389b849d9a2fec155571c9a56bcb9c3e60ed237aa"} Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.450338 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.452323 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01de3ace-e995-419f-ad27-75ee2eb0e110","Type":"ContainerStarted","Data":"f8a0c6498264abc27f52b6f5edcbe1fa57e9d7d044d3e44ea87a6ce26e9bb380"} Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.453866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8" event={"ID":"1842e17e-2c29-46ec-bdec-a6802d656e10","Type":"ContainerStarted","Data":"11bad1a4dd4079fe1f96e3c8281ffbd7aa6af278d5c660d74968d7889661dcb9"} Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.455161 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"976f8dd9-0b51-4b45-9192-9ddf098acd48","Type":"ContainerStarted","Data":"62ad13f13b8c623dee21eda3bc85c1423d8473325e34037b84a376934792841b"} Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.456454 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7sm4b" event={"ID":"953544e1-701d-4d5c-831e-77e952fc8504","Type":"ContainerStarted","Data":"fa7f51de27bf5596a185cc74c2d52ff6ff53d935399b20c30f1c69b31cce8602"} Oct 11 03:20:24 crc kubenswrapper[4754]: E1011 03:20:24.459084 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" podUID="907bf885-81ef-4141-b3f6-77da00f659f4" Oct 11 03:20:24 crc kubenswrapper[4754]: E1011 03:20:24.459447 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" podUID="0f2442f3-c008-434a-92ac-1b366deba167" Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.473131 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.697373507 podStartE2EDuration="24.473087321s" podCreationTimestamp="2025-10-11 03:20:00 +0000 UTC" firstStartedPulling="2025-10-11 03:20:01.86606779 +0000 UTC m=+849.425012575" lastFinishedPulling="2025-10-11 03:20:23.641781594 +0000 UTC m=+871.200726389" observedRunningTime="2025-10-11 03:20:24.467428709 +0000 UTC m=+872.026373514" watchObservedRunningTime="2025-10-11 03:20:24.473087321 +0000 UTC m=+872.032032106" Oct 11 03:20:24 crc kubenswrapper[4754]: I1011 03:20:24.907214 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.008181 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.078847 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvjrl\" (UniqueName: \"kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl\") pod \"ac3519a1-9e51-4225-a874-cfa333876743\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.079921 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config" (OuterVolumeSpecName: "config") pod "ac3519a1-9e51-4225-a874-cfa333876743" (UID: "ac3519a1-9e51-4225-a874-cfa333876743"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.079998 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config\") pod \"ac3519a1-9e51-4225-a874-cfa333876743\" (UID: \"ac3519a1-9e51-4225-a874-cfa333876743\") " Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.080646 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3519a1-9e51-4225-a874-cfa333876743-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.088346 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl" (OuterVolumeSpecName: "kube-api-access-nvjrl") pod "ac3519a1-9e51-4225-a874-cfa333876743" (UID: "ac3519a1-9e51-4225-a874-cfa333876743"). InnerVolumeSpecName "kube-api-access-nvjrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.152655 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.181714 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config\") pod \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.181819 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm67q\" (UniqueName: \"kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q\") pod \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.181854 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc\") pod \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\" (UID: \"379b49b8-d4fe-4608-88ba-34703cb5d6e4\") " Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.182308 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config" (OuterVolumeSpecName: "config") pod "379b49b8-d4fe-4608-88ba-34703cb5d6e4" (UID: "379b49b8-d4fe-4608-88ba-34703cb5d6e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.182889 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "379b49b8-d4fe-4608-88ba-34703cb5d6e4" (UID: "379b49b8-d4fe-4608-88ba-34703cb5d6e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.183126 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvjrl\" (UniqueName: \"kubernetes.io/projected/ac3519a1-9e51-4225-a874-cfa333876743-kube-api-access-nvjrl\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.183154 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.183167 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b49b8-d4fe-4608-88ba-34703cb5d6e4-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.185533 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q" (OuterVolumeSpecName: "kube-api-access-sm67q") pod "379b49b8-d4fe-4608-88ba-34703cb5d6e4" (UID: "379b49b8-d4fe-4608-88ba-34703cb5d6e4"). InnerVolumeSpecName "kube-api-access-sm67q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.285434 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm67q\" (UniqueName: \"kubernetes.io/projected/379b49b8-d4fe-4608-88ba-34703cb5d6e4-kube-api-access-sm67q\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.472367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" event={"ID":"379b49b8-d4fe-4608-88ba-34703cb5d6e4","Type":"ContainerDied","Data":"ea0c6d054a67f1bfb02e9d5277a5e3185968ca4a47d7834174668523d7d60805"} Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.472426 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kd5n7" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.475381 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" event={"ID":"ac3519a1-9e51-4225-a874-cfa333876743","Type":"ContainerDied","Data":"196871f00ca90ae060b25e25a3b10bcadfeb4701f07b9ba9c99cbc7065a1e806"} Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.475661 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5m5sg" Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.479875 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"daa384d9-ff26-4962-872c-68d7ffea94fe","Type":"ContainerStarted","Data":"d03f8f0ef560184b6228d1094578a299560f90e6466f9009ccfd270eba2b91d3"} Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.519658 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.529502 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5m5sg"] Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.560917 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:20:25 crc kubenswrapper[4754]: I1011 03:20:25.567245 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kd5n7"] Oct 11 03:20:27 crc kubenswrapper[4754]: I1011 03:20:27.094819 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="379b49b8-d4fe-4608-88ba-34703cb5d6e4" path="/var/lib/kubelet/pods/379b49b8-d4fe-4608-88ba-34703cb5d6e4/volumes" Oct 11 03:20:27 crc kubenswrapper[4754]: I1011 03:20:27.095616 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac3519a1-9e51-4225-a874-cfa333876743" path="/var/lib/kubelet/pods/ac3519a1-9e51-4225-a874-cfa333876743/volumes" Oct 11 03:20:27 crc kubenswrapper[4754]: I1011 03:20:27.503429 4754 generic.go:334] "Generic (PLEG): container finished" podID="01de3ace-e995-419f-ad27-75ee2eb0e110" containerID="f8a0c6498264abc27f52b6f5edcbe1fa57e9d7d044d3e44ea87a6ce26e9bb380" exitCode=0 Oct 11 03:20:27 crc kubenswrapper[4754]: I1011 03:20:27.503482 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01de3ace-e995-419f-ad27-75ee2eb0e110","Type":"ContainerDied","Data":"f8a0c6498264abc27f52b6f5edcbe1fa57e9d7d044d3e44ea87a6ce26e9bb380"} Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.516383 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"976f8dd9-0b51-4b45-9192-9ddf098acd48","Type":"ContainerStarted","Data":"e69b9f7f9abcf2f5e6a11602a13c9855ac7c757ab742a44cec0d6ecfa3113d38"} Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.517085 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.521885 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b","Type":"ContainerStarted","Data":"00e66cbc207802ace4ef1bdc8eb74d163b610643b8c2ca016631b743feee57fa"} Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.530377 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01de3ace-e995-419f-ad27-75ee2eb0e110","Type":"ContainerStarted","Data":"8e3f309cf925b3ec7368905a50889cd4897f2e0a4f476165ed628dcee73a80f9"} Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.536902 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=22.561633231 podStartE2EDuration="26.536886014s" podCreationTimestamp="2025-10-11 03:20:02 +0000 UTC" firstStartedPulling="2025-10-11 03:20:24.171839082 +0000 UTC m=+871.730783867" lastFinishedPulling="2025-10-11 03:20:28.147091865 +0000 UTC m=+875.706036650" observedRunningTime="2025-10-11 03:20:28.533011144 +0000 UTC m=+876.091955929" watchObservedRunningTime="2025-10-11 03:20:28.536886014 +0000 UTC m=+876.095830799" Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.539910 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"daa384d9-ff26-4962-872c-68d7ffea94fe","Type":"ContainerStarted","Data":"96c78b0269cd2734d46a53ae04e3a1efd5a6e7bf02d0eea12ad283624231d4b9"} Oct 11 03:20:28 crc kubenswrapper[4754]: I1011 03:20:28.563111 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.607260221 podStartE2EDuration="29.563078433s" podCreationTimestamp="2025-10-11 03:19:59 +0000 UTC" firstStartedPulling="2025-10-11 03:20:01.648690957 +0000 UTC m=+849.207635742" lastFinishedPulling="2025-10-11 03:20:23.604509169 +0000 UTC m=+871.163453954" observedRunningTime="2025-10-11 03:20:28.555148036 +0000 UTC m=+876.114092841" watchObservedRunningTime="2025-10-11 03:20:28.563078433 +0000 UTC m=+876.122023218" Oct 11 03:20:29 crc kubenswrapper[4754]: I1011 03:20:29.553253 4754 generic.go:334] "Generic (PLEG): container finished" podID="953544e1-701d-4d5c-831e-77e952fc8504" containerID="5a4ad8e002a2befc6dfbb29add8aa2eea06f2c8ca535711ddfe6f7d9677ff2e2" exitCode=0 Oct 11 03:20:29 crc kubenswrapper[4754]: I1011 03:20:29.553355 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7sm4b" event={"ID":"953544e1-701d-4d5c-831e-77e952fc8504","Type":"ContainerDied","Data":"5a4ad8e002a2befc6dfbb29add8aa2eea06f2c8ca535711ddfe6f7d9677ff2e2"} Oct 11 03:20:29 crc kubenswrapper[4754]: I1011 03:20:29.556386 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8" event={"ID":"1842e17e-2c29-46ec-bdec-a6802d656e10","Type":"ContainerStarted","Data":"4f3b79d95a4099e05d7496f59dc8fa2c15687994944ed565a740417eba636d47"} Oct 11 03:20:29 crc kubenswrapper[4754]: I1011 03:20:29.598300 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7scx8" podStartSLOduration=19.497560068 podStartE2EDuration="23.598273317s" podCreationTimestamp="2025-10-11 03:20:06 +0000 UTC" firstStartedPulling="2025-10-11 03:20:24.096485908 +0000 UTC m=+871.655430693" lastFinishedPulling="2025-10-11 03:20:28.197199147 +0000 UTC m=+875.756143942" observedRunningTime="2025-10-11 03:20:29.596813175 +0000 UTC m=+877.155757980" watchObservedRunningTime="2025-10-11 03:20:29.598273317 +0000 UTC m=+877.157218112" Oct 11 03:20:30 crc kubenswrapper[4754]: I1011 03:20:30.570682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7sm4b" event={"ID":"953544e1-701d-4d5c-831e-77e952fc8504","Type":"ContainerStarted","Data":"7f26f8599392934856bee14587bb58baa3726efb00701285fdaefc1c73364c04"} Oct 11 03:20:30 crc kubenswrapper[4754]: I1011 03:20:30.571214 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-7scx8" Oct 11 03:20:30 crc kubenswrapper[4754]: I1011 03:20:30.997072 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:30 crc kubenswrapper[4754]: I1011 03:20:30.997398 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:31 crc kubenswrapper[4754]: I1011 03:20:31.047346 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.593595 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"daa384d9-ff26-4962-872c-68d7ffea94fe","Type":"ContainerStarted","Data":"b521899aeffb1cc0a43ebf6f1e24349b6c0a6196958ab0f4be024ce017dc3347"} Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.596641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7sm4b" event={"ID":"953544e1-701d-4d5c-831e-77e952fc8504","Type":"ContainerStarted","Data":"f6beb906d0f74ae48c57152213bfd76720c6bd86fb674cf3b52365f09aabcc06"} Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.596713 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.596809 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.598894 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5fa3e4c1-2c10-4fd7-950f-380c41b8af4b","Type":"ContainerStarted","Data":"db53ea5507d915b310729229de6e038e9d0dfa60870a11654e6e626a0477b3d8"} Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.626822 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.470247365 podStartE2EDuration="26.626795724s" podCreationTimestamp="2025-10-11 03:20:06 +0000 UTC" firstStartedPulling="2025-10-11 03:20:25.281012189 +0000 UTC m=+872.839956974" lastFinishedPulling="2025-10-11 03:20:31.437560538 +0000 UTC m=+878.996505333" observedRunningTime="2025-10-11 03:20:32.618060494 +0000 UTC m=+880.177005309" watchObservedRunningTime="2025-10-11 03:20:32.626795724 +0000 UTC m=+880.185740529" Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.645543 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7sm4b" podStartSLOduration=22.692238703 podStartE2EDuration="26.645518669s" podCreationTimestamp="2025-10-11 03:20:06 +0000 UTC" firstStartedPulling="2025-10-11 03:20:24.238211888 +0000 UTC m=+871.797156673" lastFinishedPulling="2025-10-11 03:20:28.191491844 +0000 UTC m=+875.750436639" observedRunningTime="2025-10-11 03:20:32.642907695 +0000 UTC m=+880.201852520" watchObservedRunningTime="2025-10-11 03:20:32.645518669 +0000 UTC m=+880.204463454" Oct 11 03:20:32 crc kubenswrapper[4754]: I1011 03:20:32.662988 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.567828856 podStartE2EDuration="24.662946047s" podCreationTimestamp="2025-10-11 03:20:08 +0000 UTC" firstStartedPulling="2025-10-11 03:20:24.342583281 +0000 UTC m=+871.901528056" lastFinishedPulling="2025-10-11 03:20:31.437700462 +0000 UTC m=+878.996645247" observedRunningTime="2025-10-11 03:20:32.660335433 +0000 UTC m=+880.219280218" watchObservedRunningTime="2025-10-11 03:20:32.662946047 +0000 UTC m=+880.221890842" Oct 11 03:20:33 crc kubenswrapper[4754]: I1011 03:20:33.060743 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:33 crc kubenswrapper[4754]: I1011 03:20:33.071736 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:33 crc kubenswrapper[4754]: I1011 03:20:33.139514 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 11 03:20:33 crc kubenswrapper[4754]: I1011 03:20:33.612914 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerStarted","Data":"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165"} Oct 11 03:20:34 crc kubenswrapper[4754]: I1011 03:20:34.152838 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:34 crc kubenswrapper[4754]: I1011 03:20:34.228713 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:34 crc kubenswrapper[4754]: I1011 03:20:34.623366 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:34 crc kubenswrapper[4754]: I1011 03:20:34.676195 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.013524 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.052409 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.053915 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.057359 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.060113 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.097143 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.107534 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-m7h7p"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.108896 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.117842 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.126950 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.133852 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m7h7p"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovn-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232424 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rcrn\" (UniqueName: \"kubernetes.io/projected/406e66d8-4b00-4279-8acc-3e39d25b75c3-kube-api-access-6rcrn\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232489 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-combined-ca-bundle\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/406e66d8-4b00-4279-8acc-3e39d25b75c3-config\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232563 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232591 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsv89\" (UniqueName: \"kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232622 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232645 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovs-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.232719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-combined-ca-bundle\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334117 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/406e66d8-4b00-4279-8acc-3e39d25b75c3-config\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334152 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334180 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsv89\" (UniqueName: \"kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334240 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovs-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334273 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334332 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovn-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334372 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rcrn\" (UniqueName: \"kubernetes.io/projected/406e66d8-4b00-4279-8acc-3e39d25b75c3-kube-api-access-6rcrn\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.334989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/406e66d8-4b00-4279-8acc-3e39d25b75c3-config\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.335351 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovs-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.335376 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.335422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/406e66d8-4b00-4279-8acc-3e39d25b75c3-ovn-rundir\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.335696 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.336039 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.344323 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.344935 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406e66d8-4b00-4279-8acc-3e39d25b75c3-combined-ca-bundle\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.353048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsv89\" (UniqueName: \"kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89\") pod \"dnsmasq-dns-7fd796d7df-8gzgr\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.369762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rcrn\" (UniqueName: \"kubernetes.io/projected/406e66d8-4b00-4279-8acc-3e39d25b75c3-kube-api-access-6rcrn\") pod \"ovn-controller-metrics-m7h7p\" (UID: \"406e66d8-4b00-4279-8acc-3e39d25b75c3\") " pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.377439 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.424924 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.430079 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m7h7p" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.471403 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.472743 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.474588 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.490431 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.531044 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665021 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc\") pod \"907bf885-81ef-4141-b3f6-77da00f659f4\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665222 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv9w8\" (UniqueName: \"kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8\") pod \"907bf885-81ef-4141-b3f6-77da00f659f4\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665362 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config\") pod \"907bf885-81ef-4141-b3f6-77da00f659f4\" (UID: \"907bf885-81ef-4141-b3f6-77da00f659f4\") " Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665899 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.665933 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9dlx\" (UniqueName: \"kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.666033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.666054 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.667108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "907bf885-81ef-4141-b3f6-77da00f659f4" (UID: "907bf885-81ef-4141-b3f6-77da00f659f4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.667325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config" (OuterVolumeSpecName: "config") pod "907bf885-81ef-4141-b3f6-77da00f659f4" (UID: "907bf885-81ef-4141-b3f6-77da00f659f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.687688 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerStarted","Data":"ae2d41dfd9071ca61bc08bf34b22249fad90e75f35f34c5be12366fc0525a845"} Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.700437 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.700556 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pz82b" event={"ID":"907bf885-81ef-4141-b3f6-77da00f659f4","Type":"ContainerDied","Data":"5de3d6f8e9859a86212396d49e8d6064d7bca522f5bff19f383f87ffb1edae21"} Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.702248 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8" (OuterVolumeSpecName: "kube-api-access-fv9w8") pod "907bf885-81ef-4141-b3f6-77da00f659f4" (UID: "907bf885-81ef-4141-b3f6-77da00f659f4"). InnerVolumeSpecName "kube-api-access-fv9w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.767998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768065 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768223 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9dlx\" (UniqueName: \"kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768311 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv9w8\" (UniqueName: \"kubernetes.io/projected/907bf885-81ef-4141-b3f6-77da00f659f4-kube-api-access-fv9w8\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768336 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.768355 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/907bf885-81ef-4141-b3f6-77da00f659f4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.769019 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.769907 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.771018 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.771621 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.792862 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9dlx\" (UniqueName: \"kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx\") pod \"dnsmasq-dns-86db49b7ff-hplzc\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.805307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.859062 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.987699 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.989734 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.993496 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9txh9" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.993760 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.993889 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 11 03:20:35 crc kubenswrapper[4754]: I1011 03:20:35.994059 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.001907 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.041275 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076425 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-scripts\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076476 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076533 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076604 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-config\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.076652 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmwcq\" (UniqueName: \"kubernetes.io/projected/a3b76ddc-70a5-4142-8b7c-1342a716436f-kube-api-access-kmwcq\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.104483 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.111855 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pz82b"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.177822 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dtpz\" (UniqueName: \"kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz\") pod \"0f2442f3-c008-434a-92ac-1b366deba167\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178008 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc\") pod \"0f2442f3-c008-434a-92ac-1b366deba167\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178135 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config\") pod \"0f2442f3-c008-434a-92ac-1b366deba167\" (UID: \"0f2442f3-c008-434a-92ac-1b366deba167\") " Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178445 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-config\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178584 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178602 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmwcq\" (UniqueName: \"kubernetes.io/projected/a3b76ddc-70a5-4142-8b7c-1342a716436f-kube-api-access-kmwcq\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178596 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178631 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-scripts\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.178800 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.179589 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config" (OuterVolumeSpecName: "config") pod "0f2442f3-c008-434a-92ac-1b366deba167" (UID: "0f2442f3-c008-434a-92ac-1b366deba167"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.179728 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-scripts\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.179955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f2442f3-c008-434a-92ac-1b366deba167" (UID: "0f2442f3-c008-434a-92ac-1b366deba167"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.180428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.180767 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b76ddc-70a5-4142-8b7c-1342a716436f-config\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.187939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.188156 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz" (OuterVolumeSpecName: "kube-api-access-8dtpz") pod "0f2442f3-c008-434a-92ac-1b366deba167" (UID: "0f2442f3-c008-434a-92ac-1b366deba167"). InnerVolumeSpecName "kube-api-access-8dtpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.190296 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.192934 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b76ddc-70a5-4142-8b7c-1342a716436f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.195573 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmwcq\" (UniqueName: \"kubernetes.io/projected/a3b76ddc-70a5-4142-8b7c-1342a716436f-kube-api-access-kmwcq\") pod \"ovn-northd-0\" (UID: \"a3b76ddc-70a5-4142-8b7c-1342a716436f\") " pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.282297 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.282347 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dtpz\" (UniqueName: \"kubernetes.io/projected/0f2442f3-c008-434a-92ac-1b366deba167-kube-api-access-8dtpz\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.282379 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2442f3-c008-434a-92ac-1b366deba167-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.325808 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.330373 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m7h7p"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.516726 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.713955 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" event={"ID":"0f2442f3-c008-434a-92ac-1b366deba167","Type":"ContainerDied","Data":"2a6b926ee1ee03ebc5cd3a66ecab8f9c7f4109cfd1a367e901d17bb6dd18d236"} Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.714042 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vb6x4" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.715253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" event={"ID":"4f32e268-94cd-4b5b-a973-a2dfe87b9d97","Type":"ContainerStarted","Data":"4c06b8c0fb579227d6ff60d4cca8268325376ecb93cc1a840e1be468b93bfcca"} Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.718786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m7h7p" event={"ID":"406e66d8-4b00-4279-8acc-3e39d25b75c3","Type":"ContainerStarted","Data":"69d6d824ad2dd38057368a740187daeec6eaf3356fb6724db5d6fca6ce163928"} Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.718816 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m7h7p" event={"ID":"406e66d8-4b00-4279-8acc-3e39d25b75c3","Type":"ContainerStarted","Data":"e9b34122ecafcdd31e0d129fa9904b1d5b76ad66952a5b114cd754c8031790b3"} Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.721661 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" event={"ID":"8c987138-fd7a-471b-9f47-8ed30baf8547","Type":"ContainerStarted","Data":"d6d1f3bd9a05b8cfe2a2fba39f83243ffae456fd80e615f289dc8a830cccbfb6"} Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.742444 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-m7h7p" podStartSLOduration=1.7424218580000002 podStartE2EDuration="1.742421858s" podCreationTimestamp="2025-10-11 03:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:20:36.741036689 +0000 UTC m=+884.299981474" watchObservedRunningTime="2025-10-11 03:20:36.742421858 +0000 UTC m=+884.301366643" Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.846168 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.853322 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vb6x4"] Oct 11 03:20:36 crc kubenswrapper[4754]: I1011 03:20:36.861623 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 11 03:20:36 crc kubenswrapper[4754]: W1011 03:20:36.867364 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3b76ddc_70a5_4142_8b7c_1342a716436f.slice/crio-4be2a384953a11a101d711d104b3590c285b8af5a40cc076a95b8b1c89e28d96 WatchSource:0}: Error finding container 4be2a384953a11a101d711d104b3590c285b8af5a40cc076a95b8b1c89e28d96: Status 404 returned error can't find the container with id 4be2a384953a11a101d711d104b3590c285b8af5a40cc076a95b8b1c89e28d96 Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.094909 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2442f3-c008-434a-92ac-1b366deba167" path="/var/lib/kubelet/pods/0f2442f3-c008-434a-92ac-1b366deba167/volumes" Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.096091 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="907bf885-81ef-4141-b3f6-77da00f659f4" path="/var/lib/kubelet/pods/907bf885-81ef-4141-b3f6-77da00f659f4/volumes" Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.731123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe","Type":"ContainerStarted","Data":"ac61330a76375c9e2c65ee93f1f8fc74d4bd02498772576584a0f42c505da732"} Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.736377 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3b76ddc-70a5-4142-8b7c-1342a716436f","Type":"ContainerStarted","Data":"4be2a384953a11a101d711d104b3590c285b8af5a40cc076a95b8b1c89e28d96"} Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.740711 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerID="19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994" exitCode=0 Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.740786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" event={"ID":"8c987138-fd7a-471b-9f47-8ed30baf8547","Type":"ContainerDied","Data":"19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994"} Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.743419 4754 generic.go:334] "Generic (PLEG): container finished" podID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerID="529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309" exitCode=0 Oct 11 03:20:37 crc kubenswrapper[4754]: I1011 03:20:37.744112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" event={"ID":"4f32e268-94cd-4b5b-a973-a2dfe87b9d97","Type":"ContainerDied","Data":"529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309"} Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.754679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" event={"ID":"8c987138-fd7a-471b-9f47-8ed30baf8547","Type":"ContainerStarted","Data":"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d"} Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.755533 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.757385 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" event={"ID":"4f32e268-94cd-4b5b-a973-a2dfe87b9d97","Type":"ContainerStarted","Data":"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf"} Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.757644 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.759837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3b76ddc-70a5-4142-8b7c-1342a716436f","Type":"ContainerStarted","Data":"3f24c9d36ee476cee86fe806be569f75a65c74d5fb1352e48e6b2695148c6cee"} Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.759927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3b76ddc-70a5-4142-8b7c-1342a716436f","Type":"ContainerStarted","Data":"3ec034427614f679f099cb11ee05dad30c34eb9ddd94d2e6acaa2993086b6c75"} Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.760029 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.781458 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" podStartSLOduration=3.181039059 podStartE2EDuration="3.781431677s" podCreationTimestamp="2025-10-11 03:20:35 +0000 UTC" firstStartedPulling="2025-10-11 03:20:36.538920292 +0000 UTC m=+884.097865077" lastFinishedPulling="2025-10-11 03:20:37.13931291 +0000 UTC m=+884.698257695" observedRunningTime="2025-10-11 03:20:38.776301891 +0000 UTC m=+886.335246706" watchObservedRunningTime="2025-10-11 03:20:38.781431677 +0000 UTC m=+886.340376502" Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.800832 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.73208843 podStartE2EDuration="3.800810101s" podCreationTimestamp="2025-10-11 03:20:35 +0000 UTC" firstStartedPulling="2025-10-11 03:20:36.872626269 +0000 UTC m=+884.431571064" lastFinishedPulling="2025-10-11 03:20:37.94134795 +0000 UTC m=+885.500292735" observedRunningTime="2025-10-11 03:20:38.796488118 +0000 UTC m=+886.355432943" watchObservedRunningTime="2025-10-11 03:20:38.800810101 +0000 UTC m=+886.359754886" Oct 11 03:20:38 crc kubenswrapper[4754]: I1011 03:20:38.817371 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" podStartSLOduration=3.294701118 podStartE2EDuration="3.817347154s" podCreationTimestamp="2025-10-11 03:20:35 +0000 UTC" firstStartedPulling="2025-10-11 03:20:36.188780596 +0000 UTC m=+883.747725381" lastFinishedPulling="2025-10-11 03:20:36.711426632 +0000 UTC m=+884.270371417" observedRunningTime="2025-10-11 03:20:38.815407258 +0000 UTC m=+886.374352043" watchObservedRunningTime="2025-10-11 03:20:38.817347154 +0000 UTC m=+886.376291949" Oct 11 03:20:40 crc kubenswrapper[4754]: I1011 03:20:40.777903 4754 generic.go:334] "Generic (PLEG): container finished" podID="6a40ecec-4fe2-41e1-8d6c-43e7689eedfe" containerID="ac61330a76375c9e2c65ee93f1f8fc74d4bd02498772576584a0f42c505da732" exitCode=0 Oct 11 03:20:40 crc kubenswrapper[4754]: I1011 03:20:40.778005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe","Type":"ContainerDied","Data":"ac61330a76375c9e2c65ee93f1f8fc74d4bd02498772576584a0f42c505da732"} Oct 11 03:20:41 crc kubenswrapper[4754]: I1011 03:20:41.793632 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6a40ecec-4fe2-41e1-8d6c-43e7689eedfe","Type":"ContainerStarted","Data":"54e8a918b361d2527b45a9e4fbbd486df006207d536f9f242de9608150e6edc6"} Oct 11 03:20:41 crc kubenswrapper[4754]: I1011 03:20:41.852249 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371993.002573 podStartE2EDuration="43.852202872s" podCreationTimestamp="2025-10-11 03:19:58 +0000 UTC" firstStartedPulling="2025-10-11 03:20:00.390199652 +0000 UTC m=+847.949144437" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:20:41.826958311 +0000 UTC m=+889.385903136" watchObservedRunningTime="2025-10-11 03:20:41.852202872 +0000 UTC m=+889.411147677" Oct 11 03:20:42 crc kubenswrapper[4754]: I1011 03:20:42.834661 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 11 03:20:45 crc kubenswrapper[4754]: E1011 03:20:45.190183 4754 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.188:47634->38.102.83.188:46049: read tcp 38.102.83.188:47634->38.102.83.188:46049: read: connection reset by peer Oct 11 03:20:45 crc kubenswrapper[4754]: I1011 03:20:45.379218 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:45 crc kubenswrapper[4754]: I1011 03:20:45.861144 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:20:45 crc kubenswrapper[4754]: I1011 03:20:45.929564 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:45 crc kubenswrapper[4754]: I1011 03:20:45.929796 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="dnsmasq-dns" containerID="cri-o://1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf" gracePeriod=10 Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.383822 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.582136 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsv89\" (UniqueName: \"kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89\") pod \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.582484 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb\") pod \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.582572 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc\") pod \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.582659 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config\") pod \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\" (UID: \"4f32e268-94cd-4b5b-a973-a2dfe87b9d97\") " Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.588546 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89" (OuterVolumeSpecName: "kube-api-access-qsv89") pod "4f32e268-94cd-4b5b-a973-a2dfe87b9d97" (UID: "4f32e268-94cd-4b5b-a973-a2dfe87b9d97"). InnerVolumeSpecName "kube-api-access-qsv89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.621401 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config" (OuterVolumeSpecName: "config") pod "4f32e268-94cd-4b5b-a973-a2dfe87b9d97" (UID: "4f32e268-94cd-4b5b-a973-a2dfe87b9d97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.626291 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f32e268-94cd-4b5b-a973-a2dfe87b9d97" (UID: "4f32e268-94cd-4b5b-a973-a2dfe87b9d97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.630512 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f32e268-94cd-4b5b-a973-a2dfe87b9d97" (UID: "4f32e268-94cd-4b5b-a973-a2dfe87b9d97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.684649 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.684693 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.684709 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsv89\" (UniqueName: \"kubernetes.io/projected/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-kube-api-access-qsv89\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.684724 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f32e268-94cd-4b5b-a973-a2dfe87b9d97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.837081 4754 generic.go:334] "Generic (PLEG): container finished" podID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerID="1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf" exitCode=0 Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.837148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" event={"ID":"4f32e268-94cd-4b5b-a973-a2dfe87b9d97","Type":"ContainerDied","Data":"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf"} Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.837156 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.837200 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8gzgr" event={"ID":"4f32e268-94cd-4b5b-a973-a2dfe87b9d97","Type":"ContainerDied","Data":"4c06b8c0fb579227d6ff60d4cca8268325376ecb93cc1a840e1be468b93bfcca"} Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.837233 4754 scope.go:117] "RemoveContainer" containerID="1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.867413 4754 scope.go:117] "RemoveContainer" containerID="529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.885667 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.895020 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8gzgr"] Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.912225 4754 scope.go:117] "RemoveContainer" containerID="1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf" Oct 11 03:20:46 crc kubenswrapper[4754]: E1011 03:20:46.912761 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf\": container with ID starting with 1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf not found: ID does not exist" containerID="1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.912823 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf"} err="failed to get container status \"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf\": rpc error: code = NotFound desc = could not find container \"1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf\": container with ID starting with 1fe3768c9a72e49fdcaa3638797b070b881d9876164865327f7c53dbe98fe2cf not found: ID does not exist" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.912870 4754 scope.go:117] "RemoveContainer" containerID="529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309" Oct 11 03:20:46 crc kubenswrapper[4754]: E1011 03:20:46.913772 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309\": container with ID starting with 529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309 not found: ID does not exist" containerID="529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309" Oct 11 03:20:46 crc kubenswrapper[4754]: I1011 03:20:46.913819 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309"} err="failed to get container status \"529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309\": rpc error: code = NotFound desc = could not find container \"529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309\": container with ID starting with 529c232cbad0a531368de45e24608837a5da45240a32f770461ba3857c008309 not found: ID does not exist" Oct 11 03:20:47 crc kubenswrapper[4754]: I1011 03:20:47.107848 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" path="/var/lib/kubelet/pods/4f32e268-94cd-4b5b-a973-a2dfe87b9d97/volumes" Oct 11 03:20:49 crc kubenswrapper[4754]: I1011 03:20:49.618273 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 11 03:20:49 crc kubenswrapper[4754]: I1011 03:20:49.620307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 11 03:20:49 crc kubenswrapper[4754]: I1011 03:20:49.712698 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 11 03:20:49 crc kubenswrapper[4754]: I1011 03:20:49.946128 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.713210 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-cwzw8"] Oct 11 03:20:50 crc kubenswrapper[4754]: E1011 03:20:50.713805 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="init" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.713831 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="init" Oct 11 03:20:50 crc kubenswrapper[4754]: E1011 03:20:50.713861 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="dnsmasq-dns" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.713873 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="dnsmasq-dns" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.714256 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f32e268-94cd-4b5b-a973-a2dfe87b9d97" containerName="dnsmasq-dns" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.715278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.727381 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cwzw8"] Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.874204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc2c6\" (UniqueName: \"kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6\") pod \"keystone-db-create-cwzw8\" (UID: \"e37ec154-6b1a-4a5b-b622-63a4795383b7\") " pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.941293 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vmtkc"] Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.942307 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.955112 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vmtkc"] Oct 11 03:20:50 crc kubenswrapper[4754]: I1011 03:20:50.980455 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc2c6\" (UniqueName: \"kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6\") pod \"keystone-db-create-cwzw8\" (UID: \"e37ec154-6b1a-4a5b-b622-63a4795383b7\") " pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.040094 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc2c6\" (UniqueName: \"kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6\") pod \"keystone-db-create-cwzw8\" (UID: \"e37ec154-6b1a-4a5b-b622-63a4795383b7\") " pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.079484 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.104408 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcmms\" (UniqueName: \"kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms\") pod \"placement-db-create-vmtkc\" (UID: \"d24e41ae-1434-46b7-a52b-f435f48bc68e\") " pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.190876 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-s8bp9"] Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.194239 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.200173 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-s8bp9"] Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.205832 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcmms\" (UniqueName: \"kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms\") pod \"placement-db-create-vmtkc\" (UID: \"d24e41ae-1434-46b7-a52b-f435f48bc68e\") " pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.249076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcmms\" (UniqueName: \"kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms\") pod \"placement-db-create-vmtkc\" (UID: \"d24e41ae-1434-46b7-a52b-f435f48bc68e\") " pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.281491 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.313998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbkr4\" (UniqueName: \"kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4\") pod \"glance-db-create-s8bp9\" (UID: \"043dd66c-f38e-4d7b-9310-123a0a052ff0\") " pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.405127 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.415589 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbkr4\" (UniqueName: \"kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4\") pod \"glance-db-create-s8bp9\" (UID: \"043dd66c-f38e-4d7b-9310-123a0a052ff0\") " pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.445917 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbkr4\" (UniqueName: \"kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4\") pod \"glance-db-create-s8bp9\" (UID: \"043dd66c-f38e-4d7b-9310-123a0a052ff0\") " pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.517716 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.605354 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cwzw8"] Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.738557 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vmtkc"] Oct 11 03:20:51 crc kubenswrapper[4754]: W1011 03:20:51.749820 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd24e41ae_1434_46b7_a52b_f435f48bc68e.slice/crio-6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32 WatchSource:0}: Error finding container 6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32: Status 404 returned error can't find the container with id 6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32 Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.893157 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vmtkc" event={"ID":"d24e41ae-1434-46b7-a52b-f435f48bc68e","Type":"ContainerStarted","Data":"6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32"} Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.895260 4754 generic.go:334] "Generic (PLEG): container finished" podID="e37ec154-6b1a-4a5b-b622-63a4795383b7" containerID="0a0fa83857cb4fff543e37f9eb15f650c5ee77e6ebebb020e2c40dbffd363916" exitCode=0 Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.895304 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cwzw8" event={"ID":"e37ec154-6b1a-4a5b-b622-63a4795383b7","Type":"ContainerDied","Data":"0a0fa83857cb4fff543e37f9eb15f650c5ee77e6ebebb020e2c40dbffd363916"} Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.895331 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cwzw8" event={"ID":"e37ec154-6b1a-4a5b-b622-63a4795383b7","Type":"ContainerStarted","Data":"813bea5f9261580cb03a8f0d2a4d680575fd027f53564370d179d2d2546719da"} Oct 11 03:20:51 crc kubenswrapper[4754]: I1011 03:20:51.965200 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-s8bp9"] Oct 11 03:20:52 crc kubenswrapper[4754]: I1011 03:20:52.907253 4754 generic.go:334] "Generic (PLEG): container finished" podID="043dd66c-f38e-4d7b-9310-123a0a052ff0" containerID="6c2c1e6e9ec78b54748c60ebc2bf7604121f2f4a89a04b6902475f1e4de9e17c" exitCode=0 Oct 11 03:20:52 crc kubenswrapper[4754]: I1011 03:20:52.907366 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s8bp9" event={"ID":"043dd66c-f38e-4d7b-9310-123a0a052ff0","Type":"ContainerDied","Data":"6c2c1e6e9ec78b54748c60ebc2bf7604121f2f4a89a04b6902475f1e4de9e17c"} Oct 11 03:20:52 crc kubenswrapper[4754]: I1011 03:20:52.907791 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s8bp9" event={"ID":"043dd66c-f38e-4d7b-9310-123a0a052ff0","Type":"ContainerStarted","Data":"278122a5a6840de59441f0cd30c5f48f45f95b65ca744f585c86b352203e17bd"} Oct 11 03:20:52 crc kubenswrapper[4754]: I1011 03:20:52.909815 4754 generic.go:334] "Generic (PLEG): container finished" podID="d24e41ae-1434-46b7-a52b-f435f48bc68e" containerID="bf06492276a312fd0f33f83bed5c07b9c8c4268a79f0d284a5143d86a68c5bf6" exitCode=0 Oct 11 03:20:52 crc kubenswrapper[4754]: I1011 03:20:52.909895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vmtkc" event={"ID":"d24e41ae-1434-46b7-a52b-f435f48bc68e","Type":"ContainerDied","Data":"bf06492276a312fd0f33f83bed5c07b9c8c4268a79f0d284a5143d86a68c5bf6"} Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.281092 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.375021 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc2c6\" (UniqueName: \"kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6\") pod \"e37ec154-6b1a-4a5b-b622-63a4795383b7\" (UID: \"e37ec154-6b1a-4a5b-b622-63a4795383b7\") " Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.383264 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6" (OuterVolumeSpecName: "kube-api-access-cc2c6") pod "e37ec154-6b1a-4a5b-b622-63a4795383b7" (UID: "e37ec154-6b1a-4a5b-b622-63a4795383b7"). InnerVolumeSpecName "kube-api-access-cc2c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.478074 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc2c6\" (UniqueName: \"kubernetes.io/projected/e37ec154-6b1a-4a5b-b622-63a4795383b7-kube-api-access-cc2c6\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.929931 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cwzw8" Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.931853 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cwzw8" event={"ID":"e37ec154-6b1a-4a5b-b622-63a4795383b7","Type":"ContainerDied","Data":"813bea5f9261580cb03a8f0d2a4d680575fd027f53564370d179d2d2546719da"} Oct 11 03:20:53 crc kubenswrapper[4754]: I1011 03:20:53.932001 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="813bea5f9261580cb03a8f0d2a4d680575fd027f53564370d179d2d2546719da" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.410825 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.427167 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s8bp9" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.496938 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbkr4\" (UniqueName: \"kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4\") pod \"043dd66c-f38e-4d7b-9310-123a0a052ff0\" (UID: \"043dd66c-f38e-4d7b-9310-123a0a052ff0\") " Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.497663 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcmms\" (UniqueName: \"kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms\") pod \"d24e41ae-1434-46b7-a52b-f435f48bc68e\" (UID: \"d24e41ae-1434-46b7-a52b-f435f48bc68e\") " Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.504284 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4" (OuterVolumeSpecName: "kube-api-access-bbkr4") pod "043dd66c-f38e-4d7b-9310-123a0a052ff0" (UID: "043dd66c-f38e-4d7b-9310-123a0a052ff0"). InnerVolumeSpecName "kube-api-access-bbkr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.504374 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms" (OuterVolumeSpecName: "kube-api-access-zcmms") pod "d24e41ae-1434-46b7-a52b-f435f48bc68e" (UID: "d24e41ae-1434-46b7-a52b-f435f48bc68e"). InnerVolumeSpecName "kube-api-access-zcmms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.600305 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbkr4\" (UniqueName: \"kubernetes.io/projected/043dd66c-f38e-4d7b-9310-123a0a052ff0-kube-api-access-bbkr4\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.600359 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcmms\" (UniqueName: \"kubernetes.io/projected/d24e41ae-1434-46b7-a52b-f435f48bc68e-kube-api-access-zcmms\") on node \"crc\" DevicePath \"\"" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.944653 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vmtkc" event={"ID":"d24e41ae-1434-46b7-a52b-f435f48bc68e","Type":"ContainerDied","Data":"6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32"} Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.945761 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6db34462c2407f2f64f78eb1f01e877263b6c969434faf7482eaa4384a67bd32" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.944692 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vmtkc" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.947853 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-s8bp9" event={"ID":"043dd66c-f38e-4d7b-9310-123a0a052ff0","Type":"ContainerDied","Data":"278122a5a6840de59441f0cd30c5f48f45f95b65ca744f585c86b352203e17bd"} Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.947904 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="278122a5a6840de59441f0cd30c5f48f45f95b65ca744f585c86b352203e17bd" Oct 11 03:20:54 crc kubenswrapper[4754]: I1011 03:20:54.947949 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-s8bp9" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.743632 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2706-account-create-gcctp"] Oct 11 03:21:00 crc kubenswrapper[4754]: E1011 03:21:00.744631 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24e41ae-1434-46b7-a52b-f435f48bc68e" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744648 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24e41ae-1434-46b7-a52b-f435f48bc68e" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: E1011 03:21:00.744674 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="043dd66c-f38e-4d7b-9310-123a0a052ff0" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744683 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="043dd66c-f38e-4d7b-9310-123a0a052ff0" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: E1011 03:21:00.744699 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37ec154-6b1a-4a5b-b622-63a4795383b7" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744711 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37ec154-6b1a-4a5b-b622-63a4795383b7" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744913 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37ec154-6b1a-4a5b-b622-63a4795383b7" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744927 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="043dd66c-f38e-4d7b-9310-123a0a052ff0" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.744949 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24e41ae-1434-46b7-a52b-f435f48bc68e" containerName="mariadb-database-create" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.745585 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.750531 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.760551 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2706-account-create-gcctp"] Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.832193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm2l7\" (UniqueName: \"kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7\") pod \"keystone-2706-account-create-gcctp\" (UID: \"3bb448c7-4547-4717-9638-49ebda21b7c0\") " pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.933855 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm2l7\" (UniqueName: \"kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7\") pod \"keystone-2706-account-create-gcctp\" (UID: \"3bb448c7-4547-4717-9638-49ebda21b7c0\") " pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:00 crc kubenswrapper[4754]: I1011 03:21:00.957658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm2l7\" (UniqueName: \"kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7\") pod \"keystone-2706-account-create-gcctp\" (UID: \"3bb448c7-4547-4717-9638-49ebda21b7c0\") " pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.052720 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ac5e-account-create-8ph9k"] Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.054383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.057329 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.069129 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ac5e-account-create-8ph9k"] Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.074502 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.138500 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz928\" (UniqueName: \"kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928\") pod \"placement-ac5e-account-create-8ph9k\" (UID: \"f1ea6a3b-0e79-464b-bff3-67a33acca115\") " pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.243536 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz928\" (UniqueName: \"kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928\") pod \"placement-ac5e-account-create-8ph9k\" (UID: \"f1ea6a3b-0e79-464b-bff3-67a33acca115\") " pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.272246 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz928\" (UniqueName: \"kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928\") pod \"placement-ac5e-account-create-8ph9k\" (UID: \"f1ea6a3b-0e79-464b-bff3-67a33acca115\") " pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.368591 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-44df-account-create-4fm6g"] Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.369905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.371987 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.382022 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-44df-account-create-4fm6g"] Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.451645 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsxr2\" (UniqueName: \"kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2\") pod \"glance-44df-account-create-4fm6g\" (UID: \"68b62291-9634-48d3-8d23-6d12aa4c46a3\") " pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.456233 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.551537 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2706-account-create-gcctp"] Oct 11 03:21:01 crc kubenswrapper[4754]: W1011 03:21:01.551635 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bb448c7_4547_4717_9638_49ebda21b7c0.slice/crio-1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7 WatchSource:0}: Error finding container 1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7: Status 404 returned error can't find the container with id 1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7 Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.554029 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsxr2\" (UniqueName: \"kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2\") pod \"glance-44df-account-create-4fm6g\" (UID: \"68b62291-9634-48d3-8d23-6d12aa4c46a3\") " pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.573012 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsxr2\" (UniqueName: \"kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2\") pod \"glance-44df-account-create-4fm6g\" (UID: \"68b62291-9634-48d3-8d23-6d12aa4c46a3\") " pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.694460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:01 crc kubenswrapper[4754]: I1011 03:21:01.888345 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ac5e-account-create-8ph9k"] Oct 11 03:21:01 crc kubenswrapper[4754]: W1011 03:21:01.895941 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1ea6a3b_0e79_464b_bff3_67a33acca115.slice/crio-beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf WatchSource:0}: Error finding container beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf: Status 404 returned error can't find the container with id beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.042826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ac5e-account-create-8ph9k" event={"ID":"f1ea6a3b-0e79-464b-bff3-67a33acca115","Type":"ContainerStarted","Data":"beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf"} Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.045347 4754 generic.go:334] "Generic (PLEG): container finished" podID="3bb448c7-4547-4717-9638-49ebda21b7c0" containerID="23b0b42896893b37a6d3ebf8931703b08187ac6495fab267cd1dae908145f7cf" exitCode=0 Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.045371 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2706-account-create-gcctp" event={"ID":"3bb448c7-4547-4717-9638-49ebda21b7c0","Type":"ContainerDied","Data":"23b0b42896893b37a6d3ebf8931703b08187ac6495fab267cd1dae908145f7cf"} Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.045389 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2706-account-create-gcctp" event={"ID":"3bb448c7-4547-4717-9638-49ebda21b7c0","Type":"ContainerStarted","Data":"1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7"} Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.138591 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-44df-account-create-4fm6g"] Oct 11 03:21:02 crc kubenswrapper[4754]: W1011 03:21:02.178884 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b62291_9634_48d3_8d23_6d12aa4c46a3.slice/crio-fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755 WatchSource:0}: Error finding container fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755: Status 404 returned error can't find the container with id fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755 Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.294845 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-7scx8" podUID="1842e17e-2c29-46ec-bdec-a6802d656e10" containerName="ovn-controller" probeResult="failure" output=< Oct 11 03:21:02 crc kubenswrapper[4754]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 11 03:21:02 crc kubenswrapper[4754]: > Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.370405 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.386007 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7sm4b" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.598279 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7scx8-config-nb29c"] Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.599449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.602992 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.613304 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8-config-nb29c"] Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.680800 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.680857 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.680903 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.681039 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.681124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv6ht\" (UniqueName: \"kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.681154 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.782851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.782894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.782916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.782954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.783019 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv6ht\" (UniqueName: \"kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.783046 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.783273 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.783286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.783315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.784249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.785090 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.807455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv6ht\" (UniqueName: \"kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht\") pod \"ovn-controller-7scx8-config-nb29c\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:02 crc kubenswrapper[4754]: I1011 03:21:02.918516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.054876 4754 generic.go:334] "Generic (PLEG): container finished" podID="f1ea6a3b-0e79-464b-bff3-67a33acca115" containerID="d4d16df0900c21a0dc62f10b161e1ea20529f278f2175e359da129ac3e0e751a" exitCode=0 Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.055314 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ac5e-account-create-8ph9k" event={"ID":"f1ea6a3b-0e79-464b-bff3-67a33acca115","Type":"ContainerDied","Data":"d4d16df0900c21a0dc62f10b161e1ea20529f278f2175e359da129ac3e0e751a"} Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.066497 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-44df-account-create-4fm6g" event={"ID":"68b62291-9634-48d3-8d23-6d12aa4c46a3","Type":"ContainerDied","Data":"6609bdad63bd73ee2a1b36a7c111f004159ff5b59def53f9597a71049b2f118f"} Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.066352 4754 generic.go:334] "Generic (PLEG): container finished" podID="68b62291-9634-48d3-8d23-6d12aa4c46a3" containerID="6609bdad63bd73ee2a1b36a7c111f004159ff5b59def53f9597a71049b2f118f" exitCode=0 Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.066761 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-44df-account-create-4fm6g" event={"ID":"68b62291-9634-48d3-8d23-6d12aa4c46a3","Type":"ContainerStarted","Data":"fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755"} Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.382379 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8-config-nb29c"] Oct 11 03:21:03 crc kubenswrapper[4754]: W1011 03:21:03.389575 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb11f791a_2be1_434a_a010_328641e0e5d9.slice/crio-8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc WatchSource:0}: Error finding container 8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc: Status 404 returned error can't find the container with id 8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.421451 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.496745 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm2l7\" (UniqueName: \"kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7\") pod \"3bb448c7-4547-4717-9638-49ebda21b7c0\" (UID: \"3bb448c7-4547-4717-9638-49ebda21b7c0\") " Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.504053 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7" (OuterVolumeSpecName: "kube-api-access-tm2l7") pod "3bb448c7-4547-4717-9638-49ebda21b7c0" (UID: "3bb448c7-4547-4717-9638-49ebda21b7c0"). InnerVolumeSpecName "kube-api-access-tm2l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:03 crc kubenswrapper[4754]: I1011 03:21:03.599378 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm2l7\" (UniqueName: \"kubernetes.io/projected/3bb448c7-4547-4717-9638-49ebda21b7c0-kube-api-access-tm2l7\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.076800 4754 generic.go:334] "Generic (PLEG): container finished" podID="b11f791a-2be1-434a-a010-328641e0e5d9" containerID="30be761ee00c4a2c43fd00e2b0c4876fccc0fe16a7790f354cbc284fc09e4f55" exitCode=0 Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.076894 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-nb29c" event={"ID":"b11f791a-2be1-434a-a010-328641e0e5d9","Type":"ContainerDied","Data":"30be761ee00c4a2c43fd00e2b0c4876fccc0fe16a7790f354cbc284fc09e4f55"} Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.076932 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-nb29c" event={"ID":"b11f791a-2be1-434a-a010-328641e0e5d9","Type":"ContainerStarted","Data":"8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc"} Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.078764 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2706-account-create-gcctp" event={"ID":"3bb448c7-4547-4717-9638-49ebda21b7c0","Type":"ContainerDied","Data":"1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7"} Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.078833 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e54aa43f0e92cf7c4ac40884557dfd6e6edfdc5414fc0b31f64617d53b26fc7" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.078897 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2706-account-create-gcctp" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.516554 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.521914 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.617894 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz928\" (UniqueName: \"kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928\") pod \"f1ea6a3b-0e79-464b-bff3-67a33acca115\" (UID: \"f1ea6a3b-0e79-464b-bff3-67a33acca115\") " Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.617991 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsxr2\" (UniqueName: \"kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2\") pod \"68b62291-9634-48d3-8d23-6d12aa4c46a3\" (UID: \"68b62291-9634-48d3-8d23-6d12aa4c46a3\") " Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.624145 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928" (OuterVolumeSpecName: "kube-api-access-zz928") pod "f1ea6a3b-0e79-464b-bff3-67a33acca115" (UID: "f1ea6a3b-0e79-464b-bff3-67a33acca115"). InnerVolumeSpecName "kube-api-access-zz928". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.624217 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2" (OuterVolumeSpecName: "kube-api-access-xsxr2") pod "68b62291-9634-48d3-8d23-6d12aa4c46a3" (UID: "68b62291-9634-48d3-8d23-6d12aa4c46a3"). InnerVolumeSpecName "kube-api-access-xsxr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.719928 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz928\" (UniqueName: \"kubernetes.io/projected/f1ea6a3b-0e79-464b-bff3-67a33acca115-kube-api-access-zz928\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:04 crc kubenswrapper[4754]: I1011 03:21:04.719971 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsxr2\" (UniqueName: \"kubernetes.io/projected/68b62291-9634-48d3-8d23-6d12aa4c46a3-kube-api-access-xsxr2\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.111400 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ac5e-account-create-8ph9k" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.114135 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ac5e-account-create-8ph9k" event={"ID":"f1ea6a3b-0e79-464b-bff3-67a33acca115","Type":"ContainerDied","Data":"beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf"} Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.114178 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beb460ab3cdd7d282f9f9a28615cd43cf8352999a2dff394861d41ae411686cf" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.124558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-44df-account-create-4fm6g" event={"ID":"68b62291-9634-48d3-8d23-6d12aa4c46a3","Type":"ContainerDied","Data":"fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755"} Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.124622 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdefb90a097bf18ab3ac1d81fe176a7a4b2de51ad181d4403cebe01f112d9755" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.124621 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-44df-account-create-4fm6g" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.450064 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.533996 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534060 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534150 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534201 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv6ht\" (UniqueName: \"kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534250 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run\") pod \"b11f791a-2be1-434a-a010-328641e0e5d9\" (UID: \"b11f791a-2be1-434a-a010-328641e0e5d9\") " Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534756 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run" (OuterVolumeSpecName: "var-run") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534829 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.534849 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.535413 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.535774 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts" (OuterVolumeSpecName: "scripts") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.543613 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht" (OuterVolumeSpecName: "kube-api-access-fv6ht") pod "b11f791a-2be1-434a-a010-328641e0e5d9" (UID: "b11f791a-2be1-434a-a010-328641e0e5d9"). InnerVolumeSpecName "kube-api-access-fv6ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636685 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636719 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b11f791a-2be1-434a-a010-328641e0e5d9-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636732 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv6ht\" (UniqueName: \"kubernetes.io/projected/b11f791a-2be1-434a-a010-328641e0e5d9-kube-api-access-fv6ht\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636742 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636750 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:05 crc kubenswrapper[4754]: I1011 03:21:05.636759 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b11f791a-2be1-434a-a010-328641e0e5d9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.132684 4754 generic.go:334] "Generic (PLEG): container finished" podID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerID="33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165" exitCode=0 Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.132794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerDied","Data":"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165"} Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.136139 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-nb29c" event={"ID":"b11f791a-2be1-434a-a010-328641e0e5d9","Type":"ContainerDied","Data":"8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc"} Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.136179 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c336cdfa1e2ff6fd17f2b5d9b3cefd5ce21760beadadb4bb816d98ac1bc5fdc" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.136347 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-nb29c" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.524861 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-zt4j7"] Oct 11 03:21:06 crc kubenswrapper[4754]: E1011 03:21:06.525248 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ea6a3b-0e79-464b-bff3-67a33acca115" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525268 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ea6a3b-0e79-464b-bff3-67a33acca115" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: E1011 03:21:06.525297 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f791a-2be1-434a-a010-328641e0e5d9" containerName="ovn-config" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525306 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f791a-2be1-434a-a010-328641e0e5d9" containerName="ovn-config" Oct 11 03:21:06 crc kubenswrapper[4754]: E1011 03:21:06.525317 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62291-9634-48d3-8d23-6d12aa4c46a3" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525325 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62291-9634-48d3-8d23-6d12aa4c46a3" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: E1011 03:21:06.525338 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb448c7-4547-4717-9638-49ebda21b7c0" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525344 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb448c7-4547-4717-9638-49ebda21b7c0" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525501 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bb448c7-4547-4717-9638-49ebda21b7c0" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525516 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f791a-2be1-434a-a010-328641e0e5d9" containerName="ovn-config" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525529 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62291-9634-48d3-8d23-6d12aa4c46a3" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.525542 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ea6a3b-0e79-464b-bff3-67a33acca115" containerName="mariadb-account-create" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.526100 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.529082 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.529412 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ntk29" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.550453 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.550549 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.550595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s4fv\" (UniqueName: \"kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.550677 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.562298 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zt4j7"] Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.593209 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7scx8-config-nb29c"] Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.603203 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7scx8-config-nb29c"] Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.651636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.651724 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.651772 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s4fv\" (UniqueName: \"kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.651833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.670898 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.670914 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.676483 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.676612 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s4fv\" (UniqueName: \"kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv\") pod \"glance-db-sync-zt4j7\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.708755 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7scx8-config-hcrtl"] Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.709997 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.715748 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.735699 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8-config-hcrtl"] Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.752610 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.752691 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.752714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.752739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cqlk\" (UniqueName: \"kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.752977 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.753271 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.854853 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.854911 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.854978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.854998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.855026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cqlk\" (UniqueName: \"kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.855062 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.855384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.855450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.855874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.856170 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.858131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.862759 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:06 crc kubenswrapper[4754]: I1011 03:21:06.877702 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cqlk\" (UniqueName: \"kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk\") pod \"ovn-controller-7scx8-config-hcrtl\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.050542 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.094739 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11f791a-2be1-434a-a010-328641e0e5d9" path="/var/lib/kubelet/pods/b11f791a-2be1-434a-a010-328641e0e5d9/volumes" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.234170 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerID="ae2d41dfd9071ca61bc08bf34b22249fad90e75f35f34c5be12366fc0525a845" exitCode=0 Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.235096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerDied","Data":"ae2d41dfd9071ca61bc08bf34b22249fad90e75f35f34c5be12366fc0525a845"} Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.281448 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerStarted","Data":"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460"} Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.283025 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.421843 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.786352786 podStartE2EDuration="1m11.421811256s" podCreationTimestamp="2025-10-11 03:19:56 +0000 UTC" firstStartedPulling="2025-10-11 03:19:58.81300292 +0000 UTC m=+846.371947695" lastFinishedPulling="2025-10-11 03:20:31.44846138 +0000 UTC m=+879.007406165" observedRunningTime="2025-10-11 03:21:07.38834408 +0000 UTC m=+914.947288855" watchObservedRunningTime="2025-10-11 03:21:07.421811256 +0000 UTC m=+914.980756041" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.471718 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-7scx8" Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.620122 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zt4j7"] Oct 11 03:21:07 crc kubenswrapper[4754]: W1011 03:21:07.630634 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dd76760_5f2d_4c44_91f7_c8b2277df563.slice/crio-9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109 WatchSource:0}: Error finding container 9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109: Status 404 returned error can't find the container with id 9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109 Oct 11 03:21:07 crc kubenswrapper[4754]: I1011 03:21:07.654569 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7scx8-config-hcrtl"] Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.294265 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerStarted","Data":"1961cf0c165b96b95951b30ff2ff71845df237418c6f67eeebab9cc5fb0291d3"} Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.294715 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.295655 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zt4j7" event={"ID":"3dd76760-5f2d-4c44-91f7-c8b2277df563","Type":"ContainerStarted","Data":"9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109"} Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.297787 4754 generic.go:334] "Generic (PLEG): container finished" podID="711b57b4-8087-4fa0-9a0f-3eaeaf6df172" containerID="075b4884ebe3d92fcdbaac365a00f85b096e3b6f5d832b1b7bbb3613dc8f4942" exitCode=0 Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.297873 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-hcrtl" event={"ID":"711b57b4-8087-4fa0-9a0f-3eaeaf6df172","Type":"ContainerDied","Data":"075b4884ebe3d92fcdbaac365a00f85b096e3b6f5d832b1b7bbb3613dc8f4942"} Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.297901 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-hcrtl" event={"ID":"711b57b4-8087-4fa0-9a0f-3eaeaf6df172","Type":"ContainerStarted","Data":"bacb7f8897be10b903a8baa7d9bde934c638090726ee1ec5b7c5d9743dd86b1b"} Oct 11 03:21:08 crc kubenswrapper[4754]: I1011 03:21:08.331613 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371964.52319 podStartE2EDuration="1m12.331586735s" podCreationTimestamp="2025-10-11 03:19:56 +0000 UTC" firstStartedPulling="2025-10-11 03:19:58.509525437 +0000 UTC m=+846.068470222" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:08.321300281 +0000 UTC m=+915.880245086" watchObservedRunningTime="2025-10-11 03:21:08.331586735 +0000 UTC m=+915.890531520" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.649253 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728518 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728640 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run" (OuterVolumeSpecName: "var-run") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728679 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728708 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cqlk\" (UniqueName: \"kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.728935 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729071 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729114 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn\") pod \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\" (UID: \"711b57b4-8087-4fa0-9a0f-3eaeaf6df172\") " Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729188 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729624 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729649 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729663 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-var-run\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729690 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.729820 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts" (OuterVolumeSpecName: "scripts") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.737505 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk" (OuterVolumeSpecName: "kube-api-access-8cqlk") pod "711b57b4-8087-4fa0-9a0f-3eaeaf6df172" (UID: "711b57b4-8087-4fa0-9a0f-3eaeaf6df172"). InnerVolumeSpecName "kube-api-access-8cqlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.831469 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.831505 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cqlk\" (UniqueName: \"kubernetes.io/projected/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-kube-api-access-8cqlk\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:09 crc kubenswrapper[4754]: I1011 03:21:09.831521 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/711b57b4-8087-4fa0-9a0f-3eaeaf6df172-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:10 crc kubenswrapper[4754]: I1011 03:21:10.319965 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7scx8-config-hcrtl" event={"ID":"711b57b4-8087-4fa0-9a0f-3eaeaf6df172","Type":"ContainerDied","Data":"bacb7f8897be10b903a8baa7d9bde934c638090726ee1ec5b7c5d9743dd86b1b"} Oct 11 03:21:10 crc kubenswrapper[4754]: I1011 03:21:10.320018 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bacb7f8897be10b903a8baa7d9bde934c638090726ee1ec5b7c5d9743dd86b1b" Oct 11 03:21:10 crc kubenswrapper[4754]: I1011 03:21:10.320052 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7scx8-config-hcrtl" Oct 11 03:21:10 crc kubenswrapper[4754]: I1011 03:21:10.745166 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7scx8-config-hcrtl"] Oct 11 03:21:10 crc kubenswrapper[4754]: I1011 03:21:10.752325 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7scx8-config-hcrtl"] Oct 11 03:21:11 crc kubenswrapper[4754]: I1011 03:21:11.094207 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711b57b4-8087-4fa0-9a0f-3eaeaf6df172" path="/var/lib/kubelet/pods/711b57b4-8087-4fa0-9a0f-3eaeaf6df172/volumes" Oct 11 03:21:17 crc kubenswrapper[4754]: I1011 03:21:17.882878 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:21:18 crc kubenswrapper[4754]: I1011 03:21:18.279171 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.400056 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zt4j7" event={"ID":"3dd76760-5f2d-4c44-91f7-c8b2277df563","Type":"ContainerStarted","Data":"fa14052dff70eeb8ea5fc5ce7b7df7bd6af2a71fb293e179dcee756f474d9494"} Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.430561 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-zt4j7" podStartSLOduration=2.887640466 podStartE2EDuration="13.430538436s" podCreationTimestamp="2025-10-11 03:21:06 +0000 UTC" firstStartedPulling="2025-10-11 03:21:07.636318066 +0000 UTC m=+915.195262851" lastFinishedPulling="2025-10-11 03:21:18.179216046 +0000 UTC m=+925.738160821" observedRunningTime="2025-10-11 03:21:19.42578961 +0000 UTC m=+926.984734395" watchObservedRunningTime="2025-10-11 03:21:19.430538436 +0000 UTC m=+926.989483231" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.824710 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-mrg6d"] Oct 11 03:21:19 crc kubenswrapper[4754]: E1011 03:21:19.825235 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711b57b4-8087-4fa0-9a0f-3eaeaf6df172" containerName="ovn-config" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.825259 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="711b57b4-8087-4fa0-9a0f-3eaeaf6df172" containerName="ovn-config" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.825482 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="711b57b4-8087-4fa0-9a0f-3eaeaf6df172" containerName="ovn-config" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.826232 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.832511 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mrg6d"] Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.923703 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dw5l4"] Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.925145 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:19 crc kubenswrapper[4754]: I1011 03:21:19.939061 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dw5l4"] Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.001113 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jv8q\" (UniqueName: \"kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q\") pod \"cinder-db-create-mrg6d\" (UID: \"6ed0ab40-9674-4778-aaa4-8dc24d13f10a\") " pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.087241 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zjs8v"] Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.088866 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.097312 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rdhs5" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.097567 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.097725 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.100102 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.103240 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89r47\" (UniqueName: \"kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47\") pod \"barbican-db-create-dw5l4\" (UID: \"4ce57e33-81b8-4db0-aa86-cb68353f1562\") " pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.103402 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jv8q\" (UniqueName: \"kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q\") pod \"cinder-db-create-mrg6d\" (UID: \"6ed0ab40-9674-4778-aaa4-8dc24d13f10a\") " pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.110594 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zjs8v"] Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.156008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jv8q\" (UniqueName: \"kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q\") pod \"cinder-db-create-mrg6d\" (UID: \"6ed0ab40-9674-4778-aaa4-8dc24d13f10a\") " pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.207027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.207171 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q984\" (UniqueName: \"kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.207206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89r47\" (UniqueName: \"kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47\") pod \"barbican-db-create-dw5l4\" (UID: \"4ce57e33-81b8-4db0-aa86-cb68353f1562\") " pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.207303 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.210523 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6cct7"] Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.211606 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.239282 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6cct7"] Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.240117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89r47\" (UniqueName: \"kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47\") pod \"barbican-db-create-dw5l4\" (UID: \"4ce57e33-81b8-4db0-aa86-cb68353f1562\") " pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.244317 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.308517 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.308593 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcjvf\" (UniqueName: \"kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf\") pod \"neutron-db-create-6cct7\" (UID: \"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae\") " pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.308623 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.308678 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q984\" (UniqueName: \"kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.313719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.313850 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.330481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q984\" (UniqueName: \"kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984\") pod \"keystone-db-sync-zjs8v\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.409743 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcjvf\" (UniqueName: \"kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf\") pod \"neutron-db-create-6cct7\" (UID: \"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae\") " pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.423395 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.447574 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.452146 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcjvf\" (UniqueName: \"kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf\") pod \"neutron-db-create-6cct7\" (UID: \"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae\") " pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.663713 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:20 crc kubenswrapper[4754]: I1011 03:21:20.723728 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dw5l4"] Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.049599 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mrg6d"] Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.116169 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zjs8v"] Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.295353 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6cct7"] Oct 11 03:21:21 crc kubenswrapper[4754]: W1011 03:21:21.319703 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ec089c_1d8c_49ae_96ca_0bf5f36f1aae.slice/crio-a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d WatchSource:0}: Error finding container a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d: Status 404 returned error can't find the container with id a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.424600 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6cct7" event={"ID":"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae","Type":"ContainerStarted","Data":"a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d"} Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.426906 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dw5l4" event={"ID":"4ce57e33-81b8-4db0-aa86-cb68353f1562","Type":"ContainerStarted","Data":"5eb48ba0203ae5ff77ebd9dacd996dd0fae0667b22ac438d5bce4d75346d16f9"} Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.427065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dw5l4" event={"ID":"4ce57e33-81b8-4db0-aa86-cb68353f1562","Type":"ContainerStarted","Data":"18336463cf70dcf2f74b3bf87dca0032a59f2ff0bbd93fe9a0faf7f336be008c"} Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.430769 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mrg6d" event={"ID":"6ed0ab40-9674-4778-aaa4-8dc24d13f10a","Type":"ContainerStarted","Data":"930c87c78df3e06bf1acb6bc5418bf5837f7e2c4d1b2765cf7848ca78db18060"} Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.432204 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zjs8v" event={"ID":"0073f3a4-5ab0-4ee7-9793-08f96385ffd0","Type":"ContainerStarted","Data":"c1974a56abb90eaa4114089b09f4bcaf5c93408f007766384add4010a02b4001"} Oct 11 03:21:21 crc kubenswrapper[4754]: I1011 03:21:21.451250 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-dw5l4" podStartSLOduration=2.451231682 podStartE2EDuration="2.451231682s" podCreationTimestamp="2025-10-11 03:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:21.446352152 +0000 UTC m=+929.005296937" watchObservedRunningTime="2025-10-11 03:21:21.451231682 +0000 UTC m=+929.010176467" Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.448461 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6cct7" event={"ID":"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae","Type":"ContainerDied","Data":"465587eae06745a7342f9df4c5fdc149065c88c75cd088df91808ba607d2ce7f"} Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.448346 4754 generic.go:334] "Generic (PLEG): container finished" podID="a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" containerID="465587eae06745a7342f9df4c5fdc149065c88c75cd088df91808ba607d2ce7f" exitCode=0 Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.450676 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ce57e33-81b8-4db0-aa86-cb68353f1562" containerID="5eb48ba0203ae5ff77ebd9dacd996dd0fae0667b22ac438d5bce4d75346d16f9" exitCode=0 Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.450842 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dw5l4" event={"ID":"4ce57e33-81b8-4db0-aa86-cb68353f1562","Type":"ContainerDied","Data":"5eb48ba0203ae5ff77ebd9dacd996dd0fae0667b22ac438d5bce4d75346d16f9"} Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.454429 4754 generic.go:334] "Generic (PLEG): container finished" podID="6ed0ab40-9674-4778-aaa4-8dc24d13f10a" containerID="c67b92a36f1cda6583bf21d08ef40b6f604290b8f75215f26f73c91974694c3a" exitCode=0 Oct 11 03:21:22 crc kubenswrapper[4754]: I1011 03:21:22.454496 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mrg6d" event={"ID":"6ed0ab40-9674-4778-aaa4-8dc24d13f10a","Type":"ContainerDied","Data":"c67b92a36f1cda6583bf21d08ef40b6f604290b8f75215f26f73c91974694c3a"} Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.793559 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.865879 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcjvf\" (UniqueName: \"kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf\") pod \"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae\" (UID: \"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae\") " Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.870578 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf" (OuterVolumeSpecName: "kube-api-access-bcjvf") pod "a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" (UID: "a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae"). InnerVolumeSpecName "kube-api-access-bcjvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.913720 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.969022 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jv8q\" (UniqueName: \"kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q\") pod \"6ed0ab40-9674-4778-aaa4-8dc24d13f10a\" (UID: \"6ed0ab40-9674-4778-aaa4-8dc24d13f10a\") " Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.969643 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcjvf\" (UniqueName: \"kubernetes.io/projected/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae-kube-api-access-bcjvf\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.974798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q" (OuterVolumeSpecName: "kube-api-access-9jv8q") pod "6ed0ab40-9674-4778-aaa4-8dc24d13f10a" (UID: "6ed0ab40-9674-4778-aaa4-8dc24d13f10a"). InnerVolumeSpecName "kube-api-access-9jv8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:25 crc kubenswrapper[4754]: I1011 03:21:25.975476 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.070975 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89r47\" (UniqueName: \"kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47\") pod \"4ce57e33-81b8-4db0-aa86-cb68353f1562\" (UID: \"4ce57e33-81b8-4db0-aa86-cb68353f1562\") " Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.071347 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jv8q\" (UniqueName: \"kubernetes.io/projected/6ed0ab40-9674-4778-aaa4-8dc24d13f10a-kube-api-access-9jv8q\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.080084 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47" (OuterVolumeSpecName: "kube-api-access-89r47") pod "4ce57e33-81b8-4db0-aa86-cb68353f1562" (UID: "4ce57e33-81b8-4db0-aa86-cb68353f1562"). InnerVolumeSpecName "kube-api-access-89r47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.172919 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89r47\" (UniqueName: \"kubernetes.io/projected/4ce57e33-81b8-4db0-aa86-cb68353f1562-kube-api-access-89r47\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.488422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mrg6d" event={"ID":"6ed0ab40-9674-4778-aaa4-8dc24d13f10a","Type":"ContainerDied","Data":"930c87c78df3e06bf1acb6bc5418bf5837f7e2c4d1b2765cf7848ca78db18060"} Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.488471 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="930c87c78df3e06bf1acb6bc5418bf5837f7e2c4d1b2765cf7848ca78db18060" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.488543 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mrg6d" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.493263 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zjs8v" event={"ID":"0073f3a4-5ab0-4ee7-9793-08f96385ffd0","Type":"ContainerStarted","Data":"c6efcfb156d574d02111ffa339bb781cdaace237c752f7c8aa386e3be55d81f4"} Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.494542 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6cct7" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.494533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6cct7" event={"ID":"a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae","Type":"ContainerDied","Data":"a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d"} Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.494658 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a448365d535619955438f4b951e5071ac7b3edb0c5a1cfcdce6fb3ac1306840d" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.495820 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dw5l4" event={"ID":"4ce57e33-81b8-4db0-aa86-cb68353f1562","Type":"ContainerDied","Data":"18336463cf70dcf2f74b3bf87dca0032a59f2ff0bbd93fe9a0faf7f336be008c"} Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.495859 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18336463cf70dcf2f74b3bf87dca0032a59f2ff0bbd93fe9a0faf7f336be008c" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.495857 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dw5l4" Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.497100 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dd76760-5f2d-4c44-91f7-c8b2277df563" containerID="fa14052dff70eeb8ea5fc5ce7b7df7bd6af2a71fb293e179dcee756f474d9494" exitCode=0 Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.497148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zt4j7" event={"ID":"3dd76760-5f2d-4c44-91f7-c8b2277df563","Type":"ContainerDied","Data":"fa14052dff70eeb8ea5fc5ce7b7df7bd6af2a71fb293e179dcee756f474d9494"} Oct 11 03:21:26 crc kubenswrapper[4754]: I1011 03:21:26.525265 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zjs8v" podStartSLOduration=1.8802707330000001 podStartE2EDuration="6.525244645s" podCreationTimestamp="2025-10-11 03:21:20 +0000 UTC" firstStartedPulling="2025-10-11 03:21:21.141221502 +0000 UTC m=+928.700166287" lastFinishedPulling="2025-10-11 03:21:25.786195404 +0000 UTC m=+933.345140199" observedRunningTime="2025-10-11 03:21:26.524170554 +0000 UTC m=+934.083115339" watchObservedRunningTime="2025-10-11 03:21:26.525244645 +0000 UTC m=+934.084189430" Oct 11 03:21:27 crc kubenswrapper[4754]: I1011 03:21:27.981583 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.009641 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data\") pod \"3dd76760-5f2d-4c44-91f7-c8b2277df563\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.009936 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s4fv\" (UniqueName: \"kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv\") pod \"3dd76760-5f2d-4c44-91f7-c8b2277df563\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.009993 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle\") pod \"3dd76760-5f2d-4c44-91f7-c8b2277df563\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.010019 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data\") pod \"3dd76760-5f2d-4c44-91f7-c8b2277df563\" (UID: \"3dd76760-5f2d-4c44-91f7-c8b2277df563\") " Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.023107 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3dd76760-5f2d-4c44-91f7-c8b2277df563" (UID: "3dd76760-5f2d-4c44-91f7-c8b2277df563"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.023223 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv" (OuterVolumeSpecName: "kube-api-access-8s4fv") pod "3dd76760-5f2d-4c44-91f7-c8b2277df563" (UID: "3dd76760-5f2d-4c44-91f7-c8b2277df563"). InnerVolumeSpecName "kube-api-access-8s4fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.036113 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dd76760-5f2d-4c44-91f7-c8b2277df563" (UID: "3dd76760-5f2d-4c44-91f7-c8b2277df563"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.059052 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data" (OuterVolumeSpecName: "config-data") pod "3dd76760-5f2d-4c44-91f7-c8b2277df563" (UID: "3dd76760-5f2d-4c44-91f7-c8b2277df563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.111794 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.111827 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s4fv\" (UniqueName: \"kubernetes.io/projected/3dd76760-5f2d-4c44-91f7-c8b2277df563-kube-api-access-8s4fv\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.111837 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.111846 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3dd76760-5f2d-4c44-91f7-c8b2277df563-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.511184 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zt4j7" event={"ID":"3dd76760-5f2d-4c44-91f7-c8b2277df563","Type":"ContainerDied","Data":"9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109"} Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.511234 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aa46b10e52becd3f50dd7d8cb11a07ecac6e6d92e2f6818b59553f66e2e6109" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.511230 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zt4j7" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.989287 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:28 crc kubenswrapper[4754]: E1011 03:21:28.989830 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.989843 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: E1011 03:21:28.989867 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd76760-5f2d-4c44-91f7-c8b2277df563" containerName="glance-db-sync" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.989873 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd76760-5f2d-4c44-91f7-c8b2277df563" containerName="glance-db-sync" Oct 11 03:21:28 crc kubenswrapper[4754]: E1011 03:21:28.989888 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed0ab40-9674-4778-aaa4-8dc24d13f10a" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.989894 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed0ab40-9674-4778-aaa4-8dc24d13f10a" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: E1011 03:21:28.989909 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce57e33-81b8-4db0-aa86-cb68353f1562" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.989914 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce57e33-81b8-4db0-aa86-cb68353f1562" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.990062 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed0ab40-9674-4778-aaa4-8dc24d13f10a" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.990081 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd76760-5f2d-4c44-91f7-c8b2277df563" containerName="glance-db-sync" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.990094 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce57e33-81b8-4db0-aa86-cb68353f1562" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.990104 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" containerName="mariadb-database-create" Oct 11 03:21:28 crc kubenswrapper[4754]: I1011 03:21:28.990907 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.014169 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.026260 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.026315 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlllp\" (UniqueName: \"kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.026356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.026406 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.026436 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.127392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.127703 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlllp\" (UniqueName: \"kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.127849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.127977 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.128089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.128510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.128750 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.128990 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.129000 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.146417 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlllp\" (UniqueName: \"kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp\") pod \"dnsmasq-dns-54f9b7b8d9-rwp26\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.312191 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.520404 4754 generic.go:334] "Generic (PLEG): container finished" podID="0073f3a4-5ab0-4ee7-9793-08f96385ffd0" containerID="c6efcfb156d574d02111ffa339bb781cdaace237c752f7c8aa386e3be55d81f4" exitCode=0 Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.520677 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zjs8v" event={"ID":"0073f3a4-5ab0-4ee7-9793-08f96385ffd0","Type":"ContainerDied","Data":"c6efcfb156d574d02111ffa339bb781cdaace237c752f7c8aa386e3be55d81f4"} Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.737218 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:29 crc kubenswrapper[4754]: W1011 03:21:29.742317 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3671d3e_1b3d_431c_b39d_025f284dda2c.slice/crio-c4a7dc997c3df4714a60a3bdaa56ba478fda367d848b8442390849f076715425 WatchSource:0}: Error finding container c4a7dc997c3df4714a60a3bdaa56ba478fda367d848b8442390849f076715425: Status 404 returned error can't find the container with id c4a7dc997c3df4714a60a3bdaa56ba478fda367d848b8442390849f076715425 Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.967207 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5f8f-account-create-7h2dd"] Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.968841 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.970811 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 11 03:21:29 crc kubenswrapper[4754]: I1011 03:21:29.979316 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5f8f-account-create-7h2dd"] Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.044860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4fsm\" (UniqueName: \"kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm\") pod \"cinder-5f8f-account-create-7h2dd\" (UID: \"0007db7e-72b2-4e2a-b6b2-5cd94b44badf\") " pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.146426 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4fsm\" (UniqueName: \"kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm\") pod \"cinder-5f8f-account-create-7h2dd\" (UID: \"0007db7e-72b2-4e2a-b6b2-5cd94b44badf\") " pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.184065 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4fsm\" (UniqueName: \"kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm\") pod \"cinder-5f8f-account-create-7h2dd\" (UID: \"0007db7e-72b2-4e2a-b6b2-5cd94b44badf\") " pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.284295 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.528932 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerID="ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5" exitCode=0 Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.530842 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" event={"ID":"c3671d3e-1b3d-431c-b39d-025f284dda2c","Type":"ContainerDied","Data":"ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5"} Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.530883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" event={"ID":"c3671d3e-1b3d-431c-b39d-025f284dda2c","Type":"ContainerStarted","Data":"c4a7dc997c3df4714a60a3bdaa56ba478fda367d848b8442390849f076715425"} Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.721950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5f8f-account-create-7h2dd"] Oct 11 03:21:30 crc kubenswrapper[4754]: W1011 03:21:30.751325 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0007db7e_72b2_4e2a_b6b2_5cd94b44badf.slice/crio-a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a WatchSource:0}: Error finding container a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a: Status 404 returned error can't find the container with id a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.810435 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.873398 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle\") pod \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.873521 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q984\" (UniqueName: \"kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984\") pod \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.873578 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data\") pod \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\" (UID: \"0073f3a4-5ab0-4ee7-9793-08f96385ffd0\") " Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.880887 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984" (OuterVolumeSpecName: "kube-api-access-5q984") pod "0073f3a4-5ab0-4ee7-9793-08f96385ffd0" (UID: "0073f3a4-5ab0-4ee7-9793-08f96385ffd0"). InnerVolumeSpecName "kube-api-access-5q984". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.901337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0073f3a4-5ab0-4ee7-9793-08f96385ffd0" (UID: "0073f3a4-5ab0-4ee7-9793-08f96385ffd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.922905 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data" (OuterVolumeSpecName: "config-data") pod "0073f3a4-5ab0-4ee7-9793-08f96385ffd0" (UID: "0073f3a4-5ab0-4ee7-9793-08f96385ffd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.975385 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.975433 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q984\" (UniqueName: \"kubernetes.io/projected/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-kube-api-access-5q984\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:30 crc kubenswrapper[4754]: I1011 03:21:30.975447 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0073f3a4-5ab0-4ee7-9793-08f96385ffd0-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.539485 4754 generic.go:334] "Generic (PLEG): container finished" podID="0007db7e-72b2-4e2a-b6b2-5cd94b44badf" containerID="4ab2dce136d937ad284c4bf18488c37c197409308a2461e1cdcdb8ce1d692f8d" exitCode=0 Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.539719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f8f-account-create-7h2dd" event={"ID":"0007db7e-72b2-4e2a-b6b2-5cd94b44badf","Type":"ContainerDied","Data":"4ab2dce136d937ad284c4bf18488c37c197409308a2461e1cdcdb8ce1d692f8d"} Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.540744 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f8f-account-create-7h2dd" event={"ID":"0007db7e-72b2-4e2a-b6b2-5cd94b44badf","Type":"ContainerStarted","Data":"a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a"} Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.542649 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zjs8v" event={"ID":"0073f3a4-5ab0-4ee7-9793-08f96385ffd0","Type":"ContainerDied","Data":"c1974a56abb90eaa4114089b09f4bcaf5c93408f007766384add4010a02b4001"} Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.542679 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1974a56abb90eaa4114089b09f4bcaf5c93408f007766384add4010a02b4001" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.542776 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zjs8v" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.548459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" event={"ID":"c3671d3e-1b3d-431c-b39d-025f284dda2c","Type":"ContainerStarted","Data":"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce"} Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.549416 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.588311 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" podStartSLOduration=3.5882750530000003 podStartE2EDuration="3.588275053s" podCreationTimestamp="2025-10-11 03:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:31.583196258 +0000 UTC m=+939.142141043" watchObservedRunningTime="2025-10-11 03:21:31.588275053 +0000 UTC m=+939.147219838" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.784693 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.829406 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:31 crc kubenswrapper[4754]: E1011 03:21:31.833108 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0073f3a4-5ab0-4ee7-9793-08f96385ffd0" containerName="keystone-db-sync" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.833137 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0073f3a4-5ab0-4ee7-9793-08f96385ffd0" containerName="keystone-db-sync" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.833309 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0073f3a4-5ab0-4ee7-9793-08f96385ffd0" containerName="keystone-db-sync" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.834675 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.864213 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rxn7x"] Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.865600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.873397 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.873675 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rdhs5" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.873817 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.873945 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.881784 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.896518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.896839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.896922 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.896954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lzms\" (UniqueName: \"kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.897044 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.906637 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rxn7x"] Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.998920 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjllk\" (UniqueName: \"kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999557 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999585 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999616 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999637 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lzms\" (UniqueName: \"kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999683 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999731 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:31 crc kubenswrapper[4754]: I1011 03:21:31.999774 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.000658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.001234 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.002187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.002907 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.030540 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lzms\" (UniqueName: \"kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms\") pod \"dnsmasq-dns-6546db6db7-c88hj\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101462 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjllk\" (UniqueName: \"kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101503 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101528 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.101613 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.105625 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.110138 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.110825 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.111290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.112157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.132391 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.134003 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjllk\" (UniqueName: \"kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk\") pod \"keystone-bootstrap-rxn7x\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.134434 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.137579 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.138351 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.149468 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gd655"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.150586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.153741 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.153982 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-g78wt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.154125 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.156679 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.162222 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gd655"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.178021 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.192444 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202833 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202861 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d98bk\" (UniqueName: \"kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202894 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202911 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.202951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.203062 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.203080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.203121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4vc\" (UniqueName: \"kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.203138 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.204333 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.228114 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.229403 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.269020 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.310896 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311333 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4vc\" (UniqueName: \"kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2z97\" (UniqueName: \"kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311428 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311451 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311472 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d98bk\" (UniqueName: \"kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311530 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311546 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311654 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.311693 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.314056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.315412 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.323458 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.325591 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.335611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.341539 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d98bk\" (UniqueName: \"kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.349578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.350632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.362044 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.362730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4vc\" (UniqueName: \"kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.382723 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.387244 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data\") pod \"placement-db-sync-gd655\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.412916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.413009 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2z97\" (UniqueName: \"kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.413036 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.413114 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.413174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.414953 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.418397 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.428447 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.433836 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.448884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2z97\" (UniqueName: \"kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97\") pod \"dnsmasq-dns-7987f74bbc-mqmbt\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.488750 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gd655" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.534179 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.581000 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.890573 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:32 crc kubenswrapper[4754]: I1011 03:21:32.985172 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.027126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4fsm\" (UniqueName: \"kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm\") pod \"0007db7e-72b2-4e2a-b6b2-5cd94b44badf\" (UID: \"0007db7e-72b2-4e2a-b6b2-5cd94b44badf\") " Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.036177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm" (OuterVolumeSpecName: "kube-api-access-s4fsm") pod "0007db7e-72b2-4e2a-b6b2-5cd94b44badf" (UID: "0007db7e-72b2-4e2a-b6b2-5cd94b44badf"). InnerVolumeSpecName "kube-api-access-s4fsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.075228 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rxn7x"] Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.134289 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4fsm\" (UniqueName: \"kubernetes.io/projected/0007db7e-72b2-4e2a-b6b2-5cd94b44badf-kube-api-access-s4fsm\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.177426 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.197057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.204458 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gd655"] Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.570247 4754 generic.go:334] "Generic (PLEG): container finished" podID="f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" containerID="5a52f6d52a8744cfd37cf77c1179c53af183598a542dc56c3bc294d3b2cf83bf" exitCode=0 Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.570342 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" event={"ID":"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa","Type":"ContainerDied","Data":"5a52f6d52a8744cfd37cf77c1179c53af183598a542dc56c3bc294d3b2cf83bf"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.570612 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" event={"ID":"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa","Type":"ContainerStarted","Data":"b308a4f837cb078dbe67062dd2d003f8d23bf6dec96e7514ffe3d21b26302edb"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.573029 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerStarted","Data":"ae2293d633c25271b8f46e0f0782f8b363902ceeea9aaeedbbbf0bed677a806e"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.576543 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f8f-account-create-7h2dd" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.576530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f8f-account-create-7h2dd" event={"ID":"0007db7e-72b2-4e2a-b6b2-5cd94b44badf","Type":"ContainerDied","Data":"a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.576686 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a165db89b5bb0ce5dbcf7039e12f574bc93a57772849c412dc4de1988cf69b1a" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.578423 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gd655" event={"ID":"2224e5c6-2dfc-4e51-b2ee-19cc25084573","Type":"ContainerStarted","Data":"7cf5ceadec1ed118174a76d62023e85feeb977961b2661cb5daf5da42f91b461"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.579757 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerID="f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039" exitCode=0 Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.579934 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" event={"ID":"c3446bc0-2a40-4ead-926d-28ee8f8f04f0","Type":"ContainerDied","Data":"f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.579984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" event={"ID":"c3446bc0-2a40-4ead-926d-28ee8f8f04f0","Type":"ContainerStarted","Data":"db9e836a997290e94c5c3a60fc797f8d996adb9200832c2460d5017e875b8630"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.588417 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="dnsmasq-dns" containerID="cri-o://e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce" gracePeriod=10 Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.589143 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rxn7x" event={"ID":"e4f1713b-28f3-4dd9-8263-95e267fbd956","Type":"ContainerStarted","Data":"6ea1330af2899f86cb736412876e00f28b7fdee299340689c12a4dfe18c790ba"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.591336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rxn7x" event={"ID":"e4f1713b-28f3-4dd9-8263-95e267fbd956","Type":"ContainerStarted","Data":"2376668f51b547a7a7d06f5e2ed31f8bb44503fd43ff144a797b8e02b9642fd8"} Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.619218 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rxn7x" podStartSLOduration=2.619196482 podStartE2EDuration="2.619196482s" podCreationTimestamp="2025-10-11 03:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:33.612386318 +0000 UTC m=+941.171331103" watchObservedRunningTime="2025-10-11 03:21:33.619196482 +0000 UTC m=+941.178141267" Oct 11 03:21:33 crc kubenswrapper[4754]: I1011 03:21:33.959647 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.054687 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb\") pod \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.054830 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc\") pod \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.054921 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lzms\" (UniqueName: \"kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms\") pod \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.055112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb\") pod \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.055156 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config\") pod \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\" (UID: \"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.060162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms" (OuterVolumeSpecName: "kube-api-access-7lzms") pod "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" (UID: "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa"). InnerVolumeSpecName "kube-api-access-7lzms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.081045 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.081155 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config" (OuterVolumeSpecName: "config") pod "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" (UID: "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.111895 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" (UID: "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.112074 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" (UID: "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.118484 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" (UID: "f5804bd6-a993-4756-8cb7-57a6c6c8ccaa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.156813 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config\") pod \"c3671d3e-1b3d-431c-b39d-025f284dda2c\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.157025 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb\") pod \"c3671d3e-1b3d-431c-b39d-025f284dda2c\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.157096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc\") pod \"c3671d3e-1b3d-431c-b39d-025f284dda2c\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.157151 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb\") pod \"c3671d3e-1b3d-431c-b39d-025f284dda2c\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.157236 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlllp\" (UniqueName: \"kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp\") pod \"c3671d3e-1b3d-431c-b39d-025f284dda2c\" (UID: \"c3671d3e-1b3d-431c-b39d-025f284dda2c\") " Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.158035 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.158048 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.158057 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.158066 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.158074 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lzms\" (UniqueName: \"kubernetes.io/projected/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa-kube-api-access-7lzms\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.163149 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp" (OuterVolumeSpecName: "kube-api-access-mlllp") pod "c3671d3e-1b3d-431c-b39d-025f284dda2c" (UID: "c3671d3e-1b3d-431c-b39d-025f284dda2c"). InnerVolumeSpecName "kube-api-access-mlllp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.213189 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config" (OuterVolumeSpecName: "config") pod "c3671d3e-1b3d-431c-b39d-025f284dda2c" (UID: "c3671d3e-1b3d-431c-b39d-025f284dda2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.213425 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3671d3e-1b3d-431c-b39d-025f284dda2c" (UID: "c3671d3e-1b3d-431c-b39d-025f284dda2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.220351 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3671d3e-1b3d-431c-b39d-025f284dda2c" (UID: "c3671d3e-1b3d-431c-b39d-025f284dda2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.227672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3671d3e-1b3d-431c-b39d-025f284dda2c" (UID: "c3671d3e-1b3d-431c-b39d-025f284dda2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.259467 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlllp\" (UniqueName: \"kubernetes.io/projected/c3671d3e-1b3d-431c-b39d-025f284dda2c-kube-api-access-mlllp\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.259513 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.259527 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.259540 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.259552 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3671d3e-1b3d-431c-b39d-025f284dda2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.602642 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerID="e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce" exitCode=0 Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.602717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" event={"ID":"c3671d3e-1b3d-431c-b39d-025f284dda2c","Type":"ContainerDied","Data":"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce"} Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.603833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" event={"ID":"c3671d3e-1b3d-431c-b39d-025f284dda2c","Type":"ContainerDied","Data":"c4a7dc997c3df4714a60a3bdaa56ba478fda367d848b8442390849f076715425"} Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.603862 4754 scope.go:117] "RemoveContainer" containerID="e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.602785 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-rwp26" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.607282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" event={"ID":"f5804bd6-a993-4756-8cb7-57a6c6c8ccaa","Type":"ContainerDied","Data":"b308a4f837cb078dbe67062dd2d003f8d23bf6dec96e7514ffe3d21b26302edb"} Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.607388 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-c88hj" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.617523 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" event={"ID":"c3446bc0-2a40-4ead-926d-28ee8f8f04f0","Type":"ContainerStarted","Data":"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1"} Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.630122 4754 scope.go:117] "RemoveContainer" containerID="ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.666170 4754 scope.go:117] "RemoveContainer" containerID="e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce" Oct 11 03:21:34 crc kubenswrapper[4754]: E1011 03:21:34.666742 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce\": container with ID starting with e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce not found: ID does not exist" containerID="e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.669834 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce"} err="failed to get container status \"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce\": rpc error: code = NotFound desc = could not find container \"e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce\": container with ID starting with e3d3abb5f522eacfb0703f7efde0f72cc5474007d7d10a77b976d86bfce055ce not found: ID does not exist" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.670160 4754 scope.go:117] "RemoveContainer" containerID="ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5" Oct 11 03:21:34 crc kubenswrapper[4754]: E1011 03:21:34.671591 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5\": container with ID starting with ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5 not found: ID does not exist" containerID="ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.671675 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5"} err="failed to get container status \"ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5\": rpc error: code = NotFound desc = could not find container \"ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5\": container with ID starting with ec12938a22d26f8f588239eacd31f851c46e06107300ee0bd7900fc13693fcf5 not found: ID does not exist" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.671713 4754 scope.go:117] "RemoveContainer" containerID="5a52f6d52a8744cfd37cf77c1179c53af183598a542dc56c3bc294d3b2cf83bf" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.718868 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" podStartSLOduration=2.718843158 podStartE2EDuration="2.718843158s" podCreationTimestamp="2025-10-11 03:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:34.639203902 +0000 UTC m=+942.198148687" watchObservedRunningTime="2025-10-11 03:21:34.718843158 +0000 UTC m=+942.277787943" Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.728388 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.744009 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-rwp26"] Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.755911 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:34 crc kubenswrapper[4754]: I1011 03:21:34.772896 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-c88hj"] Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.121536 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" path="/var/lib/kubelet/pods/c3671d3e-1b3d-431c-b39d-025f284dda2c/volumes" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.122371 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" path="/var/lib/kubelet/pods/f5804bd6-a993-4756-8cb7-57a6c6c8ccaa/volumes" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.450296 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.486460 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vqm6s"] Oct 11 03:21:35 crc kubenswrapper[4754]: E1011 03:21:35.486777 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" containerName="init" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.486791 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" containerName="init" Oct 11 03:21:35 crc kubenswrapper[4754]: E1011 03:21:35.486806 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="dnsmasq-dns" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.486815 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="dnsmasq-dns" Oct 11 03:21:35 crc kubenswrapper[4754]: E1011 03:21:35.486829 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="init" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.486835 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="init" Oct 11 03:21:35 crc kubenswrapper[4754]: E1011 03:21:35.486857 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0007db7e-72b2-4e2a-b6b2-5cd94b44badf" containerName="mariadb-account-create" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.486862 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0007db7e-72b2-4e2a-b6b2-5cd94b44badf" containerName="mariadb-account-create" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.487023 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3671d3e-1b3d-431c-b39d-025f284dda2c" containerName="dnsmasq-dns" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.487045 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0007db7e-72b2-4e2a-b6b2-5cd94b44badf" containerName="mariadb-account-create" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.487059 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5804bd6-a993-4756-8cb7-57a6c6c8ccaa" containerName="init" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.487555 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.491520 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.491775 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.516481 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2b9z9" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.530389 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vqm6s"] Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.598880 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.598975 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.599132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7bxq\" (UniqueName: \"kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.599233 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.599419 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.599445 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.634230 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.700700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.700754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.700814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.700843 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.700933 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7bxq\" (UniqueName: \"kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.701012 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.701602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.710552 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.710871 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.714093 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.715195 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.720798 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7bxq\" (UniqueName: \"kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq\") pod \"cinder-db-sync-vqm6s\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:35 crc kubenswrapper[4754]: I1011 03:21:35.816419 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:36.278115 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vqm6s"] Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:36.642157 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqm6s" event={"ID":"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30","Type":"ContainerStarted","Data":"e051704f99021043d82867ffe2d61b6872f4662cfc596acb5c3d202344b9af71"} Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.683368 4754 generic.go:334] "Generic (PLEG): container finished" podID="e4f1713b-28f3-4dd9-8263-95e267fbd956" containerID="6ea1330af2899f86cb736412876e00f28b7fdee299340689c12a4dfe18c790ba" exitCode=0 Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.683441 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rxn7x" event={"ID":"e4f1713b-28f3-4dd9-8263-95e267fbd956","Type":"ContainerDied","Data":"6ea1330af2899f86cb736412876e00f28b7fdee299340689c12a4dfe18c790ba"} Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.928419 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-453c-account-create-kxp6w"] Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.929700 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.931715 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.946559 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-453c-account-create-kxp6w"] Oct 11 03:21:39 crc kubenswrapper[4754]: I1011 03:21:39.973204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfwq9\" (UniqueName: \"kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9\") pod \"barbican-453c-account-create-kxp6w\" (UID: \"f0101651-749d-42ea-8355-9e82199d2fe5\") " pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.074107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfwq9\" (UniqueName: \"kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9\") pod \"barbican-453c-account-create-kxp6w\" (UID: \"f0101651-749d-42ea-8355-9e82199d2fe5\") " pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.094407 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfwq9\" (UniqueName: \"kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9\") pod \"barbican-453c-account-create-kxp6w\" (UID: \"f0101651-749d-42ea-8355-9e82199d2fe5\") " pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.129776 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-336f-account-create-v5pnq"] Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.130967 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.132510 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.147279 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-336f-account-create-v5pnq"] Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.176310 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8ztj\" (UniqueName: \"kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj\") pod \"neutron-336f-account-create-v5pnq\" (UID: \"4b929a42-00e5-4d36-a15e-15b2678adb2a\") " pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.254346 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.278496 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8ztj\" (UniqueName: \"kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj\") pod \"neutron-336f-account-create-v5pnq\" (UID: \"4b929a42-00e5-4d36-a15e-15b2678adb2a\") " pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.301685 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8ztj\" (UniqueName: \"kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj\") pod \"neutron-336f-account-create-v5pnq\" (UID: \"4b929a42-00e5-4d36-a15e-15b2678adb2a\") " pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:40 crc kubenswrapper[4754]: I1011 03:21:40.461434 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:42 crc kubenswrapper[4754]: I1011 03:21:42.536800 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:21:42 crc kubenswrapper[4754]: I1011 03:21:42.601075 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:21:42 crc kubenswrapper[4754]: I1011 03:21:42.601341 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="dnsmasq-dns" containerID="cri-o://b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d" gracePeriod=10 Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.224773 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.328900 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.329118 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.329296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.329322 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.329375 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.329491 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjllk\" (UniqueName: \"kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk\") pod \"e4f1713b-28f3-4dd9-8263-95e267fbd956\" (UID: \"e4f1713b-28f3-4dd9-8263-95e267fbd956\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.336927 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts" (OuterVolumeSpecName: "scripts") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.337418 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.344056 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.344077 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk" (OuterVolumeSpecName: "kube-api-access-sjllk") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "kube-api-access-sjllk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.384987 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data" (OuterVolumeSpecName: "config-data") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.425210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-336f-account-create-v5pnq"] Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.433362 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjllk\" (UniqueName: \"kubernetes.io/projected/e4f1713b-28f3-4dd9-8263-95e267fbd956-kube-api-access-sjllk\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.433395 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.433404 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.433414 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.433423 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.440295 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4f1713b-28f3-4dd9-8263-95e267fbd956" (UID: "e4f1713b-28f3-4dd9-8263-95e267fbd956"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.444887 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.534222 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc\") pod \"8c987138-fd7a-471b-9f47-8ed30baf8547\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.534351 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb\") pod \"8c987138-fd7a-471b-9f47-8ed30baf8547\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.534483 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb\") pod \"8c987138-fd7a-471b-9f47-8ed30baf8547\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.534552 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config\") pod \"8c987138-fd7a-471b-9f47-8ed30baf8547\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.534613 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9dlx\" (UniqueName: \"kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx\") pod \"8c987138-fd7a-471b-9f47-8ed30baf8547\" (UID: \"8c987138-fd7a-471b-9f47-8ed30baf8547\") " Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.535054 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f1713b-28f3-4dd9-8263-95e267fbd956-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.538268 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx" (OuterVolumeSpecName: "kube-api-access-v9dlx") pod "8c987138-fd7a-471b-9f47-8ed30baf8547" (UID: "8c987138-fd7a-471b-9f47-8ed30baf8547"). InnerVolumeSpecName "kube-api-access-v9dlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.588738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8c987138-fd7a-471b-9f47-8ed30baf8547" (UID: "8c987138-fd7a-471b-9f47-8ed30baf8547"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.602672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c987138-fd7a-471b-9f47-8ed30baf8547" (UID: "8c987138-fd7a-471b-9f47-8ed30baf8547"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.606083 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config" (OuterVolumeSpecName: "config") pod "8c987138-fd7a-471b-9f47-8ed30baf8547" (UID: "8c987138-fd7a-471b-9f47-8ed30baf8547"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.606918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c987138-fd7a-471b-9f47-8ed30baf8547" (UID: "8c987138-fd7a-471b-9f47-8ed30baf8547"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.636619 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.636651 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.636661 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.636672 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c987138-fd7a-471b-9f47-8ed30baf8547-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.636682 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9dlx\" (UniqueName: \"kubernetes.io/projected/8c987138-fd7a-471b-9f47-8ed30baf8547-kube-api-access-v9dlx\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.702858 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-453c-account-create-kxp6w"] Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.735763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rxn7x" event={"ID":"e4f1713b-28f3-4dd9-8263-95e267fbd956","Type":"ContainerDied","Data":"2376668f51b547a7a7d06f5e2ed31f8bb44503fd43ff144a797b8e02b9642fd8"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.735797 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2376668f51b547a7a7d06f5e2ed31f8bb44503fd43ff144a797b8e02b9642fd8" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.735843 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rxn7x" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.740287 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerID="b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d" exitCode=0 Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.740383 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.741268 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" event={"ID":"8c987138-fd7a-471b-9f47-8ed30baf8547","Type":"ContainerDied","Data":"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.741315 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hplzc" event={"ID":"8c987138-fd7a-471b-9f47-8ed30baf8547","Type":"ContainerDied","Data":"d6d1f3bd9a05b8cfe2a2fba39f83243ffae456fd80e615f289dc8a830cccbfb6"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.741333 4754 scope.go:117] "RemoveContainer" containerID="b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.744459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerStarted","Data":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.747641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gd655" event={"ID":"2224e5c6-2dfc-4e51-b2ee-19cc25084573","Type":"ContainerStarted","Data":"c20ad0f1d594d79351874f09f89b7a9380d075d05cabe162e83f3bee0ed7943c"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.750236 4754 generic.go:334] "Generic (PLEG): container finished" podID="4b929a42-00e5-4d36-a15e-15b2678adb2a" containerID="928cdb0d0dcecc112a271b1cb9fc84b2f66347567fbfdacf4e3970c36ada1033" exitCode=0 Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.750275 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-336f-account-create-v5pnq" event={"ID":"4b929a42-00e5-4d36-a15e-15b2678adb2a","Type":"ContainerDied","Data":"928cdb0d0dcecc112a271b1cb9fc84b2f66347567fbfdacf4e3970c36ada1033"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.750298 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-336f-account-create-v5pnq" event={"ID":"4b929a42-00e5-4d36-a15e-15b2678adb2a","Type":"ContainerStarted","Data":"0ccf0667a0af28943fe8707d95cc0ff5afa068f6f959cb1be3554f1f62628f90"} Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.760721 4754 scope.go:117] "RemoveContainer" containerID="19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.775891 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gd655" podStartSLOduration=1.875779125 podStartE2EDuration="11.775873255s" podCreationTimestamp="2025-10-11 03:21:32 +0000 UTC" firstStartedPulling="2025-10-11 03:21:33.22475422 +0000 UTC m=+940.783699005" lastFinishedPulling="2025-10-11 03:21:43.12484836 +0000 UTC m=+950.683793135" observedRunningTime="2025-10-11 03:21:43.766116256 +0000 UTC m=+951.325061041" watchObservedRunningTime="2025-10-11 03:21:43.775873255 +0000 UTC m=+951.334818040" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.793518 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.803178 4754 scope.go:117] "RemoveContainer" containerID="b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d" Oct 11 03:21:43 crc kubenswrapper[4754]: E1011 03:21:43.803822 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d\": container with ID starting with b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d not found: ID does not exist" containerID="b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.803853 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d"} err="failed to get container status \"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d\": rpc error: code = NotFound desc = could not find container \"b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d\": container with ID starting with b164c80d37709aecc584629acc90e0b73e4e154faf17c537201fb411a7f4457d not found: ID does not exist" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.803874 4754 scope.go:117] "RemoveContainer" containerID="19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994" Oct 11 03:21:43 crc kubenswrapper[4754]: E1011 03:21:43.804258 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994\": container with ID starting with 19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994 not found: ID does not exist" containerID="19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.804281 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994"} err="failed to get container status \"19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994\": rpc error: code = NotFound desc = could not find container \"19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994\": container with ID starting with 19febcfaf679127262bd148bbebc9e4cce04e6dbd4156f9135b16488e4d6b994 not found: ID does not exist" Oct 11 03:21:43 crc kubenswrapper[4754]: I1011 03:21:43.805432 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hplzc"] Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.360115 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rxn7x"] Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.383739 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rxn7x"] Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.415116 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xxmz8"] Oct 11 03:21:44 crc kubenswrapper[4754]: E1011 03:21:44.415700 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f1713b-28f3-4dd9-8263-95e267fbd956" containerName="keystone-bootstrap" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.418995 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f1713b-28f3-4dd9-8263-95e267fbd956" containerName="keystone-bootstrap" Oct 11 03:21:44 crc kubenswrapper[4754]: E1011 03:21:44.419082 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="init" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.419127 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="init" Oct 11 03:21:44 crc kubenswrapper[4754]: E1011 03:21:44.422531 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="dnsmasq-dns" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.422550 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="dnsmasq-dns" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.422840 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" containerName="dnsmasq-dns" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.422858 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f1713b-28f3-4dd9-8263-95e267fbd956" containerName="keystone-bootstrap" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.423477 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xxmz8"] Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.423560 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.426167 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.426512 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.426560 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rdhs5" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.426625 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460017 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6gx\" (UniqueName: \"kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460049 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460110 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.460193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.562243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.563527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6gx\" (UniqueName: \"kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.563584 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.563619 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.563680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.563712 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.569212 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.570157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.570886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.577388 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.577637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.581055 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6gx\" (UniqueName: \"kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx\") pod \"keystone-bootstrap-xxmz8\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.757406 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.767472 4754 generic.go:334] "Generic (PLEG): container finished" podID="f0101651-749d-42ea-8355-9e82199d2fe5" containerID="5fcb9911414b0c21743b75a7d8084bfcac5d1ec5aa7a3070be2cd0ef4c1f37fe" exitCode=0 Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.767555 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-453c-account-create-kxp6w" event={"ID":"f0101651-749d-42ea-8355-9e82199d2fe5","Type":"ContainerDied","Data":"5fcb9911414b0c21743b75a7d8084bfcac5d1ec5aa7a3070be2cd0ef4c1f37fe"} Oct 11 03:21:44 crc kubenswrapper[4754]: I1011 03:21:44.767622 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-453c-account-create-kxp6w" event={"ID":"f0101651-749d-42ea-8355-9e82199d2fe5","Type":"ContainerStarted","Data":"802af547025b1253cc6ab1decaf521a447fce2e5bf36291324cac3d09b6180f5"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.097254 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c987138-fd7a-471b-9f47-8ed30baf8547" path="/var/lib/kubelet/pods/8c987138-fd7a-471b-9f47-8ed30baf8547/volumes" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.098197 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f1713b-28f3-4dd9-8263-95e267fbd956" path="/var/lib/kubelet/pods/e4f1713b-28f3-4dd9-8263-95e267fbd956/volumes" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.150136 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.188790 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8ztj\" (UniqueName: \"kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj\") pod \"4b929a42-00e5-4d36-a15e-15b2678adb2a\" (UID: \"4b929a42-00e5-4d36-a15e-15b2678adb2a\") " Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.197988 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj" (OuterVolumeSpecName: "kube-api-access-n8ztj") pod "4b929a42-00e5-4d36-a15e-15b2678adb2a" (UID: "4b929a42-00e5-4d36-a15e-15b2678adb2a"). InnerVolumeSpecName "kube-api-access-n8ztj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.293228 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8ztj\" (UniqueName: \"kubernetes.io/projected/4b929a42-00e5-4d36-a15e-15b2678adb2a-kube-api-access-n8ztj\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.304624 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xxmz8"] Oct 11 03:21:45 crc kubenswrapper[4754]: W1011 03:21:45.330942 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17eed71a_41c2_4333_b346_3156a8b6aaee.slice/crio-368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810 WatchSource:0}: Error finding container 368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810: Status 404 returned error can't find the container with id 368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810 Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.777622 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerStarted","Data":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.780549 4754 generic.go:334] "Generic (PLEG): container finished" podID="2224e5c6-2dfc-4e51-b2ee-19cc25084573" containerID="c20ad0f1d594d79351874f09f89b7a9380d075d05cabe162e83f3bee0ed7943c" exitCode=0 Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.780630 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gd655" event={"ID":"2224e5c6-2dfc-4e51-b2ee-19cc25084573","Type":"ContainerDied","Data":"c20ad0f1d594d79351874f09f89b7a9380d075d05cabe162e83f3bee0ed7943c"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.783668 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xxmz8" event={"ID":"17eed71a-41c2-4333-b346-3156a8b6aaee","Type":"ContainerStarted","Data":"74fdff6fd203561129ec1e5505fd464048df93295666e98d4658fdfd116b07ce"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.783693 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xxmz8" event={"ID":"17eed71a-41c2-4333-b346-3156a8b6aaee","Type":"ContainerStarted","Data":"368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.786561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-336f-account-create-v5pnq" event={"ID":"4b929a42-00e5-4d36-a15e-15b2678adb2a","Type":"ContainerDied","Data":"0ccf0667a0af28943fe8707d95cc0ff5afa068f6f959cb1be3554f1f62628f90"} Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.786608 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ccf0667a0af28943fe8707d95cc0ff5afa068f6f959cb1be3554f1f62628f90" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.786774 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-336f-account-create-v5pnq" Oct 11 03:21:45 crc kubenswrapper[4754]: I1011 03:21:45.817736 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xxmz8" podStartSLOduration=1.8177209159999999 podStartE2EDuration="1.817720916s" podCreationTimestamp="2025-10-11 03:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:21:45.814481323 +0000 UTC m=+953.373426118" watchObservedRunningTime="2025-10-11 03:21:45.817720916 +0000 UTC m=+953.376665701" Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.172944 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.221590 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfwq9\" (UniqueName: \"kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9\") pod \"f0101651-749d-42ea-8355-9e82199d2fe5\" (UID: \"f0101651-749d-42ea-8355-9e82199d2fe5\") " Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.227099 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9" (OuterVolumeSpecName: "kube-api-access-sfwq9") pod "f0101651-749d-42ea-8355-9e82199d2fe5" (UID: "f0101651-749d-42ea-8355-9e82199d2fe5"). InnerVolumeSpecName "kube-api-access-sfwq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.324166 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfwq9\" (UniqueName: \"kubernetes.io/projected/f0101651-749d-42ea-8355-9e82199d2fe5-kube-api-access-sfwq9\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.797529 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-453c-account-create-kxp6w" event={"ID":"f0101651-749d-42ea-8355-9e82199d2fe5","Type":"ContainerDied","Data":"802af547025b1253cc6ab1decaf521a447fce2e5bf36291324cac3d09b6180f5"} Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.797799 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802af547025b1253cc6ab1decaf521a447fce2e5bf36291324cac3d09b6180f5" Oct 11 03:21:46 crc kubenswrapper[4754]: I1011 03:21:46.797548 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-453c-account-create-kxp6w" Oct 11 03:21:48 crc kubenswrapper[4754]: I1011 03:21:48.815071 4754 generic.go:334] "Generic (PLEG): container finished" podID="17eed71a-41c2-4333-b346-3156a8b6aaee" containerID="74fdff6fd203561129ec1e5505fd464048df93295666e98d4658fdfd116b07ce" exitCode=0 Oct 11 03:21:48 crc kubenswrapper[4754]: I1011 03:21:48.815150 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xxmz8" event={"ID":"17eed71a-41c2-4333-b346-3156a8b6aaee","Type":"ContainerDied","Data":"74fdff6fd203561129ec1e5505fd464048df93295666e98d4658fdfd116b07ce"} Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.292765 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gd655" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.476086 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l4vc\" (UniqueName: \"kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc\") pod \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.476187 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle\") pod \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.476244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data\") pod \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.476330 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts\") pod \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.476414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs\") pod \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\" (UID: \"2224e5c6-2dfc-4e51-b2ee-19cc25084573\") " Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.477094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs" (OuterVolumeSpecName: "logs") pod "2224e5c6-2dfc-4e51-b2ee-19cc25084573" (UID: "2224e5c6-2dfc-4e51-b2ee-19cc25084573"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.482571 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc" (OuterVolumeSpecName: "kube-api-access-6l4vc") pod "2224e5c6-2dfc-4e51-b2ee-19cc25084573" (UID: "2224e5c6-2dfc-4e51-b2ee-19cc25084573"). InnerVolumeSpecName "kube-api-access-6l4vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.485526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts" (OuterVolumeSpecName: "scripts") pod "2224e5c6-2dfc-4e51-b2ee-19cc25084573" (UID: "2224e5c6-2dfc-4e51-b2ee-19cc25084573"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.508568 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2224e5c6-2dfc-4e51-b2ee-19cc25084573" (UID: "2224e5c6-2dfc-4e51-b2ee-19cc25084573"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.508598 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data" (OuterVolumeSpecName: "config-data") pod "2224e5c6-2dfc-4e51-b2ee-19cc25084573" (UID: "2224e5c6-2dfc-4e51-b2ee-19cc25084573"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.577759 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l4vc\" (UniqueName: \"kubernetes.io/projected/2224e5c6-2dfc-4e51-b2ee-19cc25084573-kube-api-access-6l4vc\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.577793 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.577802 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.577812 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224e5c6-2dfc-4e51-b2ee-19cc25084573-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.577822 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2224e5c6-2dfc-4e51-b2ee-19cc25084573-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.827645 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gd655" Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.827705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gd655" event={"ID":"2224e5c6-2dfc-4e51-b2ee-19cc25084573","Type":"ContainerDied","Data":"7cf5ceadec1ed118174a76d62023e85feeb977961b2661cb5daf5da42f91b461"} Oct 11 03:21:49 crc kubenswrapper[4754]: I1011 03:21:49.827743 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cf5ceadec1ed118174a76d62023e85feeb977961b2661cb5daf5da42f91b461" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.292447 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-jmzkc"] Oct 11 03:21:50 crc kubenswrapper[4754]: E1011 03:21:50.293084 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2224e5c6-2dfc-4e51-b2ee-19cc25084573" containerName="placement-db-sync" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293104 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2224e5c6-2dfc-4e51-b2ee-19cc25084573" containerName="placement-db-sync" Oct 11 03:21:50 crc kubenswrapper[4754]: E1011 03:21:50.293115 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b929a42-00e5-4d36-a15e-15b2678adb2a" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293122 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b929a42-00e5-4d36-a15e-15b2678adb2a" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: E1011 03:21:50.293136 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0101651-749d-42ea-8355-9e82199d2fe5" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293144 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0101651-749d-42ea-8355-9e82199d2fe5" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293309 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b929a42-00e5-4d36-a15e-15b2678adb2a" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293332 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0101651-749d-42ea-8355-9e82199d2fe5" containerName="mariadb-account-create" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293343 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2224e5c6-2dfc-4e51-b2ee-19cc25084573" containerName="placement-db-sync" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.293813 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.296113 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d2s8h" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.299196 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.310543 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jmzkc"] Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.389559 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.389658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mzbk\" (UniqueName: \"kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.389708 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.476690 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7b68f46cdd-c77ph"] Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.478546 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.482037 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.482396 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.482764 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-g78wt" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.483243 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.489385 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.489884 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-h7c6l"] Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.491787 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.493858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.494047 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.494172 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mzbk\" (UniqueName: \"kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.494215 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.494340 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.494282 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5z85h" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.500220 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.508083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.522558 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mzbk\" (UniqueName: \"kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk\") pod \"barbican-db-sync-jmzkc\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.546747 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b68f46cdd-c77ph"] Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.563296 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-h7c6l"] Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.595795 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-combined-ca-bundle\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.595843 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.595862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-config-data\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596038 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzzdv\" (UniqueName: \"kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-scripts\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596162 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtffz\" (UniqueName: \"kubernetes.io/projected/218e1b80-55c5-427d-85ef-b35d891f69a0-kube-api-access-qtffz\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-public-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218e1b80-55c5-427d-85ef-b35d891f69a0-logs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.596406 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-internal-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.614670 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698233 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-combined-ca-bundle\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698343 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-config-data\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698430 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzzdv\" (UniqueName: \"kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-scripts\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698522 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtffz\" (UniqueName: \"kubernetes.io/projected/218e1b80-55c5-427d-85ef-b35d891f69a0-kube-api-access-qtffz\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698546 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-public-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698576 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218e1b80-55c5-427d-85ef-b35d891f69a0-logs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.698615 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-internal-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.699249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218e1b80-55c5-427d-85ef-b35d891f69a0-logs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.705362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-scripts\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.706006 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-public-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.706640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-combined-ca-bundle\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.708348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-config-data\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.709071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.709651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218e1b80-55c5-427d-85ef-b35d891f69a0-internal-tls-certs\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.713239 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.716495 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzzdv\" (UniqueName: \"kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv\") pod \"neutron-db-sync-h7c6l\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.720812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtffz\" (UniqueName: \"kubernetes.io/projected/218e1b80-55c5-427d-85ef-b35d891f69a0-kube-api-access-qtffz\") pod \"placement-7b68f46cdd-c77ph\" (UID: \"218e1b80-55c5-427d-85ef-b35d891f69a0\") " pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.889840 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:21:50 crc kubenswrapper[4754]: I1011 03:21:50.898157 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:22:00 crc kubenswrapper[4754]: I1011 03:22:00.736905 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:22:00 crc kubenswrapper[4754]: I1011 03:22:00.737625 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:22:02 crc kubenswrapper[4754]: E1011 03:22:02.666942 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 11 03:22:02 crc kubenswrapper[4754]: E1011 03:22:02.667239 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7bxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vqm6s_openstack(3ed9167b-ca5f-4fe5-b362-2d2bdef58d30): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:22:02 crc kubenswrapper[4754]: E1011 03:22:02.668518 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vqm6s" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.776232 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r6gx\" (UniqueName: \"kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940545 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940588 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940654 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940753 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.940786 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys\") pod \"17eed71a-41c2-4333-b346-3156a8b6aaee\" (UID: \"17eed71a-41c2-4333-b346-3156a8b6aaee\") " Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.946495 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx" (OuterVolumeSpecName: "kube-api-access-5r6gx") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "kube-api-access-5r6gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.948835 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts" (OuterVolumeSpecName: "scripts") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.948907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.949438 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.956100 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerStarted","Data":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.957938 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xxmz8" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.958047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xxmz8" event={"ID":"17eed71a-41c2-4333-b346-3156a8b6aaee","Type":"ContainerDied","Data":"368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810"} Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.958068 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="368c36b815dc06e507002ce67cc63b75cd20488fde931bb7bb370b8cc5df3810" Oct 11 03:22:02 crc kubenswrapper[4754]: E1011 03:22:02.962789 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vqm6s" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.980180 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:02 crc kubenswrapper[4754]: I1011 03:22:02.997063 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data" (OuterVolumeSpecName: "config-data") pod "17eed71a-41c2-4333-b346-3156a8b6aaee" (UID: "17eed71a-41c2-4333-b346-3156a8b6aaee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043090 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043119 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043129 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043141 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r6gx\" (UniqueName: \"kubernetes.io/projected/17eed71a-41c2-4333-b346-3156a8b6aaee-kube-api-access-5r6gx\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043149 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.043157 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eed71a-41c2-4333-b346-3156a8b6aaee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.114741 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jmzkc"] Oct 11 03:22:03 crc kubenswrapper[4754]: W1011 03:22:03.121879 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93df894a_1420_42c8_9352_5acd38935208.slice/crio-22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481 WatchSource:0}: Error finding container 22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481: Status 404 returned error can't find the container with id 22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481 Oct 11 03:22:03 crc kubenswrapper[4754]: W1011 03:22:03.169219 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78bf9306_4051_4919_9276_e89b67330f04.slice/crio-76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1 WatchSource:0}: Error finding container 76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1: Status 404 returned error can't find the container with id 76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1 Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.171553 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-h7c6l"] Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.229691 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b68f46cdd-c77ph"] Oct 11 03:22:03 crc kubenswrapper[4754]: W1011 03:22:03.252293 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod218e1b80_55c5_427d_85ef_b35d891f69a0.slice/crio-6af12ad5a2db815ebef54cf3a9adef1f00da5b3d733df512496d335edb14687a WatchSource:0}: Error finding container 6af12ad5a2db815ebef54cf3a9adef1f00da5b3d733df512496d335edb14687a: Status 404 returned error can't find the container with id 6af12ad5a2db815ebef54cf3a9adef1f00da5b3d733df512496d335edb14687a Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.879872 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-748c895d7d-pnbk5"] Oct 11 03:22:03 crc kubenswrapper[4754]: E1011 03:22:03.880628 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eed71a-41c2-4333-b346-3156a8b6aaee" containerName="keystone-bootstrap" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.880647 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eed71a-41c2-4333-b346-3156a8b6aaee" containerName="keystone-bootstrap" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.880849 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="17eed71a-41c2-4333-b346-3156a8b6aaee" containerName="keystone-bootstrap" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.882018 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.884936 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.885242 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.887737 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rdhs5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.887750 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.888039 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.888161 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.907245 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-748c895d7d-pnbk5"] Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957138 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-public-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957189 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-internal-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957219 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-credential-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957282 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-combined-ca-bundle\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-scripts\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957380 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-config-data\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957441 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-fernet-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.957566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68prg\" (UniqueName: \"kubernetes.io/projected/5ab66911-8a46-403c-9a45-e70b9cc46b60-kube-api-access-68prg\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.973421 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7c6l" event={"ID":"78bf9306-4051-4919-9276-e89b67330f04","Type":"ContainerStarted","Data":"a3b0d8781811253873f50d0a476fcd7e6e914deed9e39ea0d34ae12bac2a316c"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.973463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7c6l" event={"ID":"78bf9306-4051-4919-9276-e89b67330f04","Type":"ContainerStarted","Data":"76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.976004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b68f46cdd-c77ph" event={"ID":"218e1b80-55c5-427d-85ef-b35d891f69a0","Type":"ContainerStarted","Data":"7b3d0a8f8ddc2e1010847ca13abb8a0fb5bf54f632f214eb1e54786d3e43e3bd"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.976029 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b68f46cdd-c77ph" event={"ID":"218e1b80-55c5-427d-85ef-b35d891f69a0","Type":"ContainerStarted","Data":"efa9a9d7f2d37c4caef03be19148ac68e767e0d0a7cf08389f39b563b73332b1"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.976041 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b68f46cdd-c77ph" event={"ID":"218e1b80-55c5-427d-85ef-b35d891f69a0","Type":"ContainerStarted","Data":"6af12ad5a2db815ebef54cf3a9adef1f00da5b3d733df512496d335edb14687a"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.976796 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.976825 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.978425 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmzkc" event={"ID":"93df894a-1420-42c8-9352-5acd38935208","Type":"ContainerStarted","Data":"22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481"} Oct 11 03:22:03 crc kubenswrapper[4754]: I1011 03:22:03.990774 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-h7c6l" podStartSLOduration=13.990762115999999 podStartE2EDuration="13.990762116s" podCreationTimestamp="2025-10-11 03:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:03.986180295 +0000 UTC m=+971.545125080" watchObservedRunningTime="2025-10-11 03:22:03.990762116 +0000 UTC m=+971.549706901" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.010137 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7b68f46cdd-c77ph" podStartSLOduration=14.010118899 podStartE2EDuration="14.010118899s" podCreationTimestamp="2025-10-11 03:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:04.004345054 +0000 UTC m=+971.563289849" watchObservedRunningTime="2025-10-11 03:22:04.010118899 +0000 UTC m=+971.569063684" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68prg\" (UniqueName: \"kubernetes.io/projected/5ab66911-8a46-403c-9a45-e70b9cc46b60-kube-api-access-68prg\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058526 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-public-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058545 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-internal-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058575 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-credential-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-combined-ca-bundle\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058663 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-scripts\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-config-data\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.058707 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-fernet-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.064821 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-public-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.065209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-scripts\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.065243 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-config-data\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.065486 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-combined-ca-bundle\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.068618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-fernet-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.069534 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-credential-keys\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.073144 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68prg\" (UniqueName: \"kubernetes.io/projected/5ab66911-8a46-403c-9a45-e70b9cc46b60-kube-api-access-68prg\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.083726 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ab66911-8a46-403c-9a45-e70b9cc46b60-internal-tls-certs\") pod \"keystone-748c895d7d-pnbk5\" (UID: \"5ab66911-8a46-403c-9a45-e70b9cc46b60\") " pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.203918 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.662879 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-748c895d7d-pnbk5"] Oct 11 03:22:04 crc kubenswrapper[4754]: W1011 03:22:04.672621 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ab66911_8a46_403c_9a45_e70b9cc46b60.slice/crio-10db1d1b22c26a5be98bf33892e16dade15cb88facc1824670b3d74378c60c76 WatchSource:0}: Error finding container 10db1d1b22c26a5be98bf33892e16dade15cb88facc1824670b3d74378c60c76: Status 404 returned error can't find the container with id 10db1d1b22c26a5be98bf33892e16dade15cb88facc1824670b3d74378c60c76 Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.988269 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-748c895d7d-pnbk5" event={"ID":"5ab66911-8a46-403c-9a45-e70b9cc46b60","Type":"ContainerStarted","Data":"f9388c090b7f880b31c7d4911daa0ed16a633ccc640e5dcfdebc6cb57e4eefa3"} Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.988603 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-748c895d7d-pnbk5" event={"ID":"5ab66911-8a46-403c-9a45-e70b9cc46b60","Type":"ContainerStarted","Data":"10db1d1b22c26a5be98bf33892e16dade15cb88facc1824670b3d74378c60c76"} Oct 11 03:22:04 crc kubenswrapper[4754]: I1011 03:22:04.988886 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:05 crc kubenswrapper[4754]: I1011 03:22:05.010767 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-748c895d7d-pnbk5" podStartSLOduration=2.010744835 podStartE2EDuration="2.010744835s" podCreationTimestamp="2025-10-11 03:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:05.004765424 +0000 UTC m=+972.563710209" watchObservedRunningTime="2025-10-11 03:22:05.010744835 +0000 UTC m=+972.569689640" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.055513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerStarted","Data":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.055618 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-central-agent" containerID="cri-o://fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" gracePeriod=30 Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.056544 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.055795 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="sg-core" containerID="cri-o://1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" gracePeriod=30 Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.055742 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-notification-agent" containerID="cri-o://8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" gracePeriod=30 Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.056253 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="proxy-httpd" containerID="cri-o://a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" gracePeriod=30 Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.066715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmzkc" event={"ID":"93df894a-1420-42c8-9352-5acd38935208","Type":"ContainerStarted","Data":"bef964d39dce76ff5239b9ab3edc3c38ae1a78e707cacdf50fcdd413b2f8b188"} Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.083215 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.611289138 podStartE2EDuration="40.083196011s" podCreationTimestamp="2025-10-11 03:21:32 +0000 UTC" firstStartedPulling="2025-10-11 03:21:33.218277335 +0000 UTC m=+940.777222120" lastFinishedPulling="2025-10-11 03:22:11.690184218 +0000 UTC m=+979.249128993" observedRunningTime="2025-10-11 03:22:12.08245726 +0000 UTC m=+979.641402095" watchObservedRunningTime="2025-10-11 03:22:12.083196011 +0000 UTC m=+979.642140796" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.105700 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-jmzkc" podStartSLOduration=13.642037578 podStartE2EDuration="22.105680865s" podCreationTimestamp="2025-10-11 03:21:50 +0000 UTC" firstStartedPulling="2025-10-11 03:22:03.123884033 +0000 UTC m=+970.682828818" lastFinishedPulling="2025-10-11 03:22:11.58752731 +0000 UTC m=+979.146472105" observedRunningTime="2025-10-11 03:22:12.10051585 +0000 UTC m=+979.659460655" watchObservedRunningTime="2025-10-11 03:22:12.105680865 +0000 UTC m=+979.664625660" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.712719 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.910903 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.910965 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911018 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911077 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d98bk\" (UniqueName: \"kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911107 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911247 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data\") pod \"19a3ac3d-3c77-407a-b898-be69261632f7\" (UID: \"19a3ac3d-3c77-407a-b898-be69261632f7\") " Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.911912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.912111 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.918271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts" (OuterVolumeSpecName: "scripts") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.918324 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk" (OuterVolumeSpecName: "kube-api-access-d98bk") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "kube-api-access-d98bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:12 crc kubenswrapper[4754]: I1011 03:22:12.956108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.007982 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013003 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013025 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013034 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013045 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013055 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d98bk\" (UniqueName: \"kubernetes.io/projected/19a3ac3d-3c77-407a-b898-be69261632f7-kube-api-access-d98bk\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.013064 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19a3ac3d-3c77-407a-b898-be69261632f7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.033586 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data" (OuterVolumeSpecName: "config-data") pod "19a3ac3d-3c77-407a-b898-be69261632f7" (UID: "19a3ac3d-3c77-407a-b898-be69261632f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.081497 4754 generic.go:334] "Generic (PLEG): container finished" podID="19a3ac3d-3c77-407a-b898-be69261632f7" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" exitCode=0 Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.082811 4754 generic.go:334] "Generic (PLEG): container finished" podID="19a3ac3d-3c77-407a-b898-be69261632f7" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" exitCode=2 Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.083077 4754 generic.go:334] "Generic (PLEG): container finished" podID="19a3ac3d-3c77-407a-b898-be69261632f7" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" exitCode=0 Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.083214 4754 generic.go:334] "Generic (PLEG): container finished" podID="19a3ac3d-3c77-407a-b898-be69261632f7" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" exitCode=0 Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.081555 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.081579 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerDied","Data":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.083410 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerDied","Data":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.083432 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerDied","Data":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.083913 4754 scope.go:117] "RemoveContainer" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.096699 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerDied","Data":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.096739 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19a3ac3d-3c77-407a-b898-be69261632f7","Type":"ContainerDied","Data":"ae2293d633c25271b8f46e0f0782f8b363902ceeea9aaeedbbbf0bed677a806e"} Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.104489 4754 scope.go:117] "RemoveContainer" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.115230 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a3ac3d-3c77-407a-b898-be69261632f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.127152 4754 scope.go:117] "RemoveContainer" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.155040 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.163400 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.167836 4754 scope.go:117] "RemoveContainer" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.192495 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.193352 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-central-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.193442 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-central-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.193588 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-notification-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.193662 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-notification-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.193752 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="proxy-httpd" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.193817 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="proxy-httpd" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.194572 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="sg-core" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.194669 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="sg-core" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.194953 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="proxy-httpd" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.195065 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="sg-core" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.195152 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-central-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.195240 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" containerName="ceilometer-notification-agent" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.197444 4754 scope.go:117] "RemoveContainer" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.197858 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.198150 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": container with ID starting with a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c not found: ID does not exist" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198185 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} err="failed to get container status \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": rpc error: code = NotFound desc = could not find container \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": container with ID starting with a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198213 4754 scope.go:117] "RemoveContainer" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.198584 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": container with ID starting with 1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee not found: ID does not exist" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198615 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} err="failed to get container status \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": rpc error: code = NotFound desc = could not find container \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": container with ID starting with 1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198642 4754 scope.go:117] "RemoveContainer" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.198831 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": container with ID starting with 8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5 not found: ID does not exist" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198849 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} err="failed to get container status \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": rpc error: code = NotFound desc = could not find container \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": container with ID starting with 8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.198861 4754 scope.go:117] "RemoveContainer" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: E1011 03:22:13.199101 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": container with ID starting with fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953 not found: ID does not exist" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.199197 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} err="failed to get container status \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": rpc error: code = NotFound desc = could not find container \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": container with ID starting with fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.199304 4754 scope.go:117] "RemoveContainer" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.202968 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203180 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} err="failed to get container status \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": rpc error: code = NotFound desc = could not find container \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": container with ID starting with a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203230 4754 scope.go:117] "RemoveContainer" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203481 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} err="failed to get container status \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": rpc error: code = NotFound desc = could not find container \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": container with ID starting with 1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203518 4754 scope.go:117] "RemoveContainer" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203775 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} err="failed to get container status \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": rpc error: code = NotFound desc = could not find container \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": container with ID starting with 8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.203797 4754 scope.go:117] "RemoveContainer" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204006 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} err="failed to get container status \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": rpc error: code = NotFound desc = could not find container \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": container with ID starting with fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204030 4754 scope.go:117] "RemoveContainer" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204409 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} err="failed to get container status \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": rpc error: code = NotFound desc = could not find container \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": container with ID starting with a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204431 4754 scope.go:117] "RemoveContainer" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204609 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} err="failed to get container status \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": rpc error: code = NotFound desc = could not find container \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": container with ID starting with 1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204629 4754 scope.go:117] "RemoveContainer" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204817 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} err="failed to get container status \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": rpc error: code = NotFound desc = could not find container \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": container with ID starting with 8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.204850 4754 scope.go:117] "RemoveContainer" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205073 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} err="failed to get container status \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": rpc error: code = NotFound desc = could not find container \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": container with ID starting with fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205093 4754 scope.go:117] "RemoveContainer" containerID="a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205308 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c"} err="failed to get container status \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": rpc error: code = NotFound desc = could not find container \"a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c\": container with ID starting with a4c8b0566df49593d0d726575dfc14995ceba3a2b8cc65b404798a5a3346b84c not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205327 4754 scope.go:117] "RemoveContainer" containerID="1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205511 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee"} err="failed to get container status \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": rpc error: code = NotFound desc = could not find container \"1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee\": container with ID starting with 1b5e8030d72521bbf437ed019b7eef3a039fd59547164fc26eaf705f513558ee not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205537 4754 scope.go:117] "RemoveContainer" containerID="8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205717 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5"} err="failed to get container status \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": rpc error: code = NotFound desc = could not find container \"8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5\": container with ID starting with 8c7e335d3d0d2c57f7c4061982c4bcd794aabccd9dfcf64e8cb3143db36bf4c5 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205744 4754 scope.go:117] "RemoveContainer" containerID="fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.205960 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953"} err="failed to get container status \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": rpc error: code = NotFound desc = could not find container \"fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953\": container with ID starting with fdf084376261467db1f786106b552fb00fb7f8ae66a43cf8f33b7a5257695953 not found: ID does not exist" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217418 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n22f4\" (UniqueName: \"kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217493 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217521 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217567 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.217796 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.233247 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.234159 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n22f4\" (UniqueName: \"kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320131 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320169 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.320269 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.325040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.328135 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.328757 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.329019 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.333123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.337613 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.348418 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n22f4\" (UniqueName: \"kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4\") pod \"ceilometer-0\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.525433 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:13 crc kubenswrapper[4754]: I1011 03:22:13.956464 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:13 crc kubenswrapper[4754]: W1011 03:22:13.961557 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7acd45c_a0be_46e0_9834_f23183f0673a.slice/crio-6616035a579ee8fd099fb5fa89db8ebdc3c306ec02cd8318271d40295068ed2f WatchSource:0}: Error finding container 6616035a579ee8fd099fb5fa89db8ebdc3c306ec02cd8318271d40295068ed2f: Status 404 returned error can't find the container with id 6616035a579ee8fd099fb5fa89db8ebdc3c306ec02cd8318271d40295068ed2f Oct 11 03:22:14 crc kubenswrapper[4754]: I1011 03:22:14.092210 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerStarted","Data":"6616035a579ee8fd099fb5fa89db8ebdc3c306ec02cd8318271d40295068ed2f"} Oct 11 03:22:15 crc kubenswrapper[4754]: I1011 03:22:15.094101 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a3ac3d-3c77-407a-b898-be69261632f7" path="/var/lib/kubelet/pods/19a3ac3d-3c77-407a-b898-be69261632f7/volumes" Oct 11 03:22:15 crc kubenswrapper[4754]: I1011 03:22:15.107623 4754 generic.go:334] "Generic (PLEG): container finished" podID="93df894a-1420-42c8-9352-5acd38935208" containerID="bef964d39dce76ff5239b9ab3edc3c38ae1a78e707cacdf50fcdd413b2f8b188" exitCode=0 Oct 11 03:22:15 crc kubenswrapper[4754]: I1011 03:22:15.107687 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmzkc" event={"ID":"93df894a-1420-42c8-9352-5acd38935208","Type":"ContainerDied","Data":"bef964d39dce76ff5239b9ab3edc3c38ae1a78e707cacdf50fcdd413b2f8b188"} Oct 11 03:22:15 crc kubenswrapper[4754]: I1011 03:22:15.110140 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerStarted","Data":"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c"} Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.120365 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerStarted","Data":"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b"} Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.120971 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerStarted","Data":"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f"} Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.423617 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.581152 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mzbk\" (UniqueName: \"kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk\") pod \"93df894a-1420-42c8-9352-5acd38935208\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.581254 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle\") pod \"93df894a-1420-42c8-9352-5acd38935208\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.581290 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data\") pod \"93df894a-1420-42c8-9352-5acd38935208\" (UID: \"93df894a-1420-42c8-9352-5acd38935208\") " Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.589090 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "93df894a-1420-42c8-9352-5acd38935208" (UID: "93df894a-1420-42c8-9352-5acd38935208"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.589141 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk" (OuterVolumeSpecName: "kube-api-access-7mzbk") pod "93df894a-1420-42c8-9352-5acd38935208" (UID: "93df894a-1420-42c8-9352-5acd38935208"). InnerVolumeSpecName "kube-api-access-7mzbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.610957 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93df894a-1420-42c8-9352-5acd38935208" (UID: "93df894a-1420-42c8-9352-5acd38935208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.683461 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mzbk\" (UniqueName: \"kubernetes.io/projected/93df894a-1420-42c8-9352-5acd38935208-kube-api-access-7mzbk\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.683490 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:16 crc kubenswrapper[4754]: I1011 03:22:16.683499 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93df894a-1420-42c8-9352-5acd38935208-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.153539 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqm6s" event={"ID":"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30","Type":"ContainerStarted","Data":"89781785b76dbf12528636ee8c09f05422be24e2f63fa4e5b664991c246401e2"} Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.168475 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmzkc" event={"ID":"93df894a-1420-42c8-9352-5acd38935208","Type":"ContainerDied","Data":"22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481"} Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.168524 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a23ab3709547e9414be1bc62a50c21e39d12c4314cf16e0ed9a96d569d4481" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.168619 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmzkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.197554 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vqm6s" podStartSLOduration=2.781688931 podStartE2EDuration="42.197534572s" podCreationTimestamp="2025-10-11 03:21:35 +0000 UTC" firstStartedPulling="2025-10-11 03:21:36.288652469 +0000 UTC m=+943.847597244" lastFinishedPulling="2025-10-11 03:22:15.7044981 +0000 UTC m=+983.263442885" observedRunningTime="2025-10-11 03:22:17.194594749 +0000 UTC m=+984.753539544" watchObservedRunningTime="2025-10-11 03:22:17.197534572 +0000 UTC m=+984.756479367" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.402929 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-66d4bc859f-nl9q7"] Oct 11 03:22:17 crc kubenswrapper[4754]: E1011 03:22:17.403310 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93df894a-1420-42c8-9352-5acd38935208" containerName="barbican-db-sync" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.403326 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="93df894a-1420-42c8-9352-5acd38935208" containerName="barbican-db-sync" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.403500 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="93df894a-1420-42c8-9352-5acd38935208" containerName="barbican-db-sync" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.404388 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.426210 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f89475ccd-9rw9s"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.427549 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.430775 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d2s8h" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.430998 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.431128 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.431263 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.443309 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66d4bc859f-nl9q7"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.463734 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f89475ccd-9rw9s"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502682 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-logs\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-combined-ca-bundle\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502765 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223cdab5-c653-4b55-a14f-c6206ee53f32-logs\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502789 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502814 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j428r\" (UniqueName: \"kubernetes.io/projected/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-kube-api-access-j428r\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data-custom\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502875 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-combined-ca-bundle\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502904 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbsqq\" (UniqueName: \"kubernetes.io/projected/223cdab5-c653-4b55-a14f-c6206ee53f32-kube-api-access-xbsqq\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502920 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data-custom\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.502977 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.526241 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.528165 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.546191 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605374 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbsqq\" (UniqueName: \"kubernetes.io/projected/223cdab5-c653-4b55-a14f-c6206ee53f32-kube-api-access-xbsqq\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605440 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data-custom\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605496 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605523 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605610 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-logs\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605722 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-combined-ca-bundle\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605777 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223cdab5-c653-4b55-a14f-c6206ee53f32-logs\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605837 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcqbb\" (UniqueName: \"kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605886 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605937 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j428r\" (UniqueName: \"kubernetes.io/projected/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-kube-api-access-j428r\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.605990 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.606085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data-custom\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.606142 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-combined-ca-bundle\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.606346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223cdab5-c653-4b55-a14f-c6206ee53f32-logs\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.609348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-logs\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.613377 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.616206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data-custom\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.620940 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-config-data\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.621105 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-config-data-custom\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.621392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223cdab5-c653-4b55-a14f-c6206ee53f32-combined-ca-bundle\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.621898 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-combined-ca-bundle\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.630755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j428r\" (UniqueName: \"kubernetes.io/projected/eb6c5d2c-fd76-44e2-a563-09b7dab7c352-kube-api-access-j428r\") pod \"barbican-keystone-listener-7f89475ccd-9rw9s\" (UID: \"eb6c5d2c-fd76-44e2-a563-09b7dab7c352\") " pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.640179 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbsqq\" (UniqueName: \"kubernetes.io/projected/223cdab5-c653-4b55-a14f-c6206ee53f32-kube-api-access-xbsqq\") pod \"barbican-worker-66d4bc859f-nl9q7\" (UID: \"223cdab5-c653-4b55-a14f-c6206ee53f32\") " pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.682347 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.685295 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.690434 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.712637 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.713614 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.713645 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.713709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.713728 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcqbb\" (UniqueName: \"kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.713761 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.714784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.724485 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.725045 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.725117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.725286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66d4bc859f-nl9q7" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.737875 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcqbb\" (UniqueName: \"kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb\") pod \"dnsmasq-dns-699df9757c-sqztm\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.755843 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.815007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.815058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.815097 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.815153 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w78ll\" (UniqueName: \"kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.815185 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.853845 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.916955 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w78ll\" (UniqueName: \"kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.917257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.917318 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.917377 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.917417 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.918796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.926141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.933061 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w78ll\" (UniqueName: \"kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.935741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:17 crc kubenswrapper[4754]: I1011 03:22:17.940256 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data\") pod \"barbican-api-85464d6c8b-t4kkc\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.129769 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.195284 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerStarted","Data":"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4"} Oct 11 03:22:18 crc kubenswrapper[4754]: W1011 03:22:18.195310 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1663c50_eb8f_4a31_9f7b_f8d720848e56.slice/crio-30b1f961b5810a0e47a90e3556452871e7cb7afa790bc7d90f6ac9e557050b8f WatchSource:0}: Error finding container 30b1f961b5810a0e47a90e3556452871e7cb7afa790bc7d90f6ac9e557050b8f: Status 404 returned error can't find the container with id 30b1f961b5810a0e47a90e3556452871e7cb7afa790bc7d90f6ac9e557050b8f Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.195766 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.204892 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.228255 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.562032895 podStartE2EDuration="5.228240403s" podCreationTimestamp="2025-10-11 03:22:13 +0000 UTC" firstStartedPulling="2025-10-11 03:22:13.963939564 +0000 UTC m=+981.522884349" lastFinishedPulling="2025-10-11 03:22:17.630147072 +0000 UTC m=+985.189091857" observedRunningTime="2025-10-11 03:22:18.227475431 +0000 UTC m=+985.786420226" watchObservedRunningTime="2025-10-11 03:22:18.228240403 +0000 UTC m=+985.787185188" Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.255638 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66d4bc859f-nl9q7"] Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.325694 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f89475ccd-9rw9s"] Oct 11 03:22:18 crc kubenswrapper[4754]: I1011 03:22:18.784920 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.218682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" event={"ID":"eb6c5d2c-fd76-44e2-a563-09b7dab7c352","Type":"ContainerStarted","Data":"d5d3ec9e7f12ca7fc99b481441a0bd9355db8c1cd563b738a143eecff104da0a"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.220626 4754 generic.go:334] "Generic (PLEG): container finished" podID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerID="2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6" exitCode=0 Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.220684 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699df9757c-sqztm" event={"ID":"f1663c50-eb8f-4a31-9f7b-f8d720848e56","Type":"ContainerDied","Data":"2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.220709 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699df9757c-sqztm" event={"ID":"f1663c50-eb8f-4a31-9f7b-f8d720848e56","Type":"ContainerStarted","Data":"30b1f961b5810a0e47a90e3556452871e7cb7afa790bc7d90f6ac9e557050b8f"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.249251 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66d4bc859f-nl9q7" event={"ID":"223cdab5-c653-4b55-a14f-c6206ee53f32","Type":"ContainerStarted","Data":"d97050261f38a85d4650c06250968ea26843fbdff801ae17324cdf4638a79e47"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.261337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerStarted","Data":"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.261730 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerStarted","Data":"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.261787 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerStarted","Data":"6462fed4fd9fd5d93dfd4ce21e3020f324fcab4d617532200c0600ce02ed716f"} Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.261858 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.261921 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:19 crc kubenswrapper[4754]: I1011 03:22:19.282963 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85464d6c8b-t4kkc" podStartSLOduration=2.282945652 podStartE2EDuration="2.282945652s" podCreationTimestamp="2025-10-11 03:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:19.279623028 +0000 UTC m=+986.838567813" watchObservedRunningTime="2025-10-11 03:22:19.282945652 +0000 UTC m=+986.841890437" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.279027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699df9757c-sqztm" event={"ID":"f1663c50-eb8f-4a31-9f7b-f8d720848e56","Type":"ContainerStarted","Data":"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52"} Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.279546 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.302225 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-699df9757c-sqztm" podStartSLOduration=3.302208521 podStartE2EDuration="3.302208521s" podCreationTimestamp="2025-10-11 03:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:20.298478636 +0000 UTC m=+987.857423421" watchObservedRunningTime="2025-10-11 03:22:20.302208521 +0000 UTC m=+987.861153306" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.611191 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6658bb6dd-v68dl"] Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.612578 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.615109 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.615294 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.656521 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6658bb6dd-v68dl"] Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-internal-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-public-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data-custom\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l87c6\" (UniqueName: \"kubernetes.io/projected/17d5b752-8e1d-4fe0-ad28-fc3073f96157-kube-api-access-l87c6\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d5b752-8e1d-4fe0-ad28-fc3073f96157-logs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684596 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.684728 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-combined-ca-bundle\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786271 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l87c6\" (UniqueName: \"kubernetes.io/projected/17d5b752-8e1d-4fe0-ad28-fc3073f96157-kube-api-access-l87c6\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786344 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d5b752-8e1d-4fe0-ad28-fc3073f96157-logs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786446 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-combined-ca-bundle\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786501 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-internal-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786773 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-public-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786807 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data-custom\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.786991 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d5b752-8e1d-4fe0-ad28-fc3073f96157-logs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.791727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-internal-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.792173 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.793099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-public-tls-certs\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.795897 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-config-data-custom\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.796233 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d5b752-8e1d-4fe0-ad28-fc3073f96157-combined-ca-bundle\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.806328 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l87c6\" (UniqueName: \"kubernetes.io/projected/17d5b752-8e1d-4fe0-ad28-fc3073f96157-kube-api-access-l87c6\") pod \"barbican-api-6658bb6dd-v68dl\" (UID: \"17d5b752-8e1d-4fe0-ad28-fc3073f96157\") " pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:20 crc kubenswrapper[4754]: I1011 03:22:20.928607 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.287462 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66d4bc859f-nl9q7" event={"ID":"223cdab5-c653-4b55-a14f-c6206ee53f32","Type":"ContainerStarted","Data":"ef772528e3b95fef691f9961a16ef94e74158af690fa6a0fe55ebe83c897c261"} Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.287735 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66d4bc859f-nl9q7" event={"ID":"223cdab5-c653-4b55-a14f-c6206ee53f32","Type":"ContainerStarted","Data":"633b161040b9968a2cc8d211bcc6690db68da1d8c5672449fc0194c8f3e06d61"} Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.291213 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" event={"ID":"eb6c5d2c-fd76-44e2-a563-09b7dab7c352","Type":"ContainerStarted","Data":"0f4bd28ea35dfb9cf878a0fae3b32e713a50ece9c58aa75baf13a0f1738ff6bb"} Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.291266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" event={"ID":"eb6c5d2c-fd76-44e2-a563-09b7dab7c352","Type":"ContainerStarted","Data":"296ba922eee64233b42efda386c51876f2a06fced13d3ed21d53ef460413ac79"} Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.305775 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-66d4bc859f-nl9q7" podStartSLOduration=2.134311028 podStartE2EDuration="4.305757057s" podCreationTimestamp="2025-10-11 03:22:17 +0000 UTC" firstStartedPulling="2025-10-11 03:22:18.27666627 +0000 UTC m=+985.835611055" lastFinishedPulling="2025-10-11 03:22:20.448112299 +0000 UTC m=+988.007057084" observedRunningTime="2025-10-11 03:22:21.302240757 +0000 UTC m=+988.861185542" watchObservedRunningTime="2025-10-11 03:22:21.305757057 +0000 UTC m=+988.864701832" Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.321875 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f89475ccd-9rw9s" podStartSLOduration=2.205573008 podStartE2EDuration="4.321857231s" podCreationTimestamp="2025-10-11 03:22:17 +0000 UTC" firstStartedPulling="2025-10-11 03:22:18.333262357 +0000 UTC m=+985.892207142" lastFinishedPulling="2025-10-11 03:22:20.44954657 +0000 UTC m=+988.008491365" observedRunningTime="2025-10-11 03:22:21.319608978 +0000 UTC m=+988.878553763" watchObservedRunningTime="2025-10-11 03:22:21.321857231 +0000 UTC m=+988.880802016" Oct 11 03:22:21 crc kubenswrapper[4754]: I1011 03:22:21.474525 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6658bb6dd-v68dl"] Oct 11 03:22:21 crc kubenswrapper[4754]: W1011 03:22:21.480086 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17d5b752_8e1d_4fe0_ad28_fc3073f96157.slice/crio-43fcf4139dc0844666a12ede3e2b72540d4e6ced60031b5668dfdceb824ec84a WatchSource:0}: Error finding container 43fcf4139dc0844666a12ede3e2b72540d4e6ced60031b5668dfdceb824ec84a: Status 404 returned error can't find the container with id 43fcf4139dc0844666a12ede3e2b72540d4e6ced60031b5668dfdceb824ec84a Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.307201 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6658bb6dd-v68dl" event={"ID":"17d5b752-8e1d-4fe0-ad28-fc3073f96157","Type":"ContainerStarted","Data":"50168a23186fcb3ca1b40001945a9796d3e0aa079887d6fe54941def5e5b7315"} Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.307456 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.307467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6658bb6dd-v68dl" event={"ID":"17d5b752-8e1d-4fe0-ad28-fc3073f96157","Type":"ContainerStarted","Data":"a72e0c623c9bd50495293afe9eda62d2203ed2d03ddd7f28d2df896ca9307fa1"} Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.307480 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6658bb6dd-v68dl" event={"ID":"17d5b752-8e1d-4fe0-ad28-fc3073f96157","Type":"ContainerStarted","Data":"43fcf4139dc0844666a12ede3e2b72540d4e6ced60031b5668dfdceb824ec84a"} Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.307498 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.338230 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6658bb6dd-v68dl" podStartSLOduration=2.338211487 podStartE2EDuration="2.338211487s" podCreationTimestamp="2025-10-11 03:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:22.333084632 +0000 UTC m=+989.892029417" watchObservedRunningTime="2025-10-11 03:22:22.338211487 +0000 UTC m=+989.897156272" Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.484005 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:22:22 crc kubenswrapper[4754]: I1011 03:22:22.490646 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b68f46cdd-c77ph" Oct 11 03:22:23 crc kubenswrapper[4754]: I1011 03:22:23.320044 4754 generic.go:334] "Generic (PLEG): container finished" podID="78bf9306-4051-4919-9276-e89b67330f04" containerID="a3b0d8781811253873f50d0a476fcd7e6e914deed9e39ea0d34ae12bac2a316c" exitCode=0 Oct 11 03:22:23 crc kubenswrapper[4754]: I1011 03:22:23.320096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7c6l" event={"ID":"78bf9306-4051-4919-9276-e89b67330f04","Type":"ContainerDied","Data":"a3b0d8781811253873f50d0a476fcd7e6e914deed9e39ea0d34ae12bac2a316c"} Oct 11 03:22:23 crc kubenswrapper[4754]: I1011 03:22:23.321742 4754 generic.go:334] "Generic (PLEG): container finished" podID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" containerID="89781785b76dbf12528636ee8c09f05422be24e2f63fa4e5b664991c246401e2" exitCode=0 Oct 11 03:22:23 crc kubenswrapper[4754]: I1011 03:22:23.321858 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqm6s" event={"ID":"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30","Type":"ContainerDied","Data":"89781785b76dbf12528636ee8c09f05422be24e2f63fa4e5b664991c246401e2"} Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.770170 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.782925 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.863986 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864086 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzzdv\" (UniqueName: \"kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv\") pod \"78bf9306-4051-4919-9276-e89b67330f04\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864240 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864266 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864543 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864631 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7bxq\" (UniqueName: \"kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq\") pod \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\" (UID: \"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864734 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config\") pod \"78bf9306-4051-4919-9276-e89b67330f04\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.864800 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") pod \"78bf9306-4051-4919-9276-e89b67330f04\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.867437 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.870525 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts" (OuterVolumeSpecName: "scripts") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.876356 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.880203 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq" (OuterVolumeSpecName: "kube-api-access-z7bxq") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "kube-api-access-z7bxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.880363 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.890357 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv" (OuterVolumeSpecName: "kube-api-access-tzzdv") pod "78bf9306-4051-4919-9276-e89b67330f04" (UID: "78bf9306-4051-4919-9276-e89b67330f04"). InnerVolumeSpecName "kube-api-access-tzzdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: E1011 03:22:24.902259 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle podName:78bf9306-4051-4919-9276-e89b67330f04 nodeName:}" failed. No retries permitted until 2025-10-11 03:22:25.402226277 +0000 UTC m=+992.961171062 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle") pod "78bf9306-4051-4919-9276-e89b67330f04" (UID: "78bf9306-4051-4919-9276-e89b67330f04") : error deleting /var/lib/kubelet/pods/78bf9306-4051-4919-9276-e89b67330f04/volume-subpaths: remove /var/lib/kubelet/pods/78bf9306-4051-4919-9276-e89b67330f04/volume-subpaths: no such file or directory Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.909075 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config" (OuterVolumeSpecName: "config") pod "78bf9306-4051-4919-9276-e89b67330f04" (UID: "78bf9306-4051-4919-9276-e89b67330f04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.924471 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.930058 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data" (OuterVolumeSpecName: "config-data") pod "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" (UID: "3ed9167b-ca5f-4fe5-b362-2d2bdef58d30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967415 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzzdv\" (UniqueName: \"kubernetes.io/projected/78bf9306-4051-4919-9276-e89b67330f04-kube-api-access-tzzdv\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967502 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967524 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967543 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967566 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967585 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7bxq\" (UniqueName: \"kubernetes.io/projected/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-kube-api-access-z7bxq\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967603 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:24 crc kubenswrapper[4754]: I1011 03:22:24.967620 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.340134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7c6l" event={"ID":"78bf9306-4051-4919-9276-e89b67330f04","Type":"ContainerDied","Data":"76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1"} Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.340186 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76911a9026ab2c91c4964954a6b932e508acd73513cbffd99fbb3c8b8d1440c1" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.340271 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7c6l" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.343039 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqm6s" event={"ID":"3ed9167b-ca5f-4fe5-b362-2d2bdef58d30","Type":"ContainerDied","Data":"e051704f99021043d82867ffe2d61b6872f4662cfc596acb5c3d202344b9af71"} Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.343066 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e051704f99021043d82867ffe2d61b6872f4662cfc596acb5c3d202344b9af71" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.343116 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqm6s" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.489784 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") pod \"78bf9306-4051-4919-9276-e89b67330f04\" (UID: \"78bf9306-4051-4919-9276-e89b67330f04\") " Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.497103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78bf9306-4051-4919-9276-e89b67330f04" (UID: "78bf9306-4051-4919-9276-e89b67330f04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.592416 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78bf9306-4051-4919-9276-e89b67330f04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.687539 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.687814 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-699df9757c-sqztm" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="dnsmasq-dns" containerID="cri-o://19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52" gracePeriod=10 Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.689200 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.709443 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:25 crc kubenswrapper[4754]: E1011 03:22:25.710177 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78bf9306-4051-4919-9276-e89b67330f04" containerName="neutron-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.710201 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="78bf9306-4051-4919-9276-e89b67330f04" containerName="neutron-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: E1011 03:22:25.710230 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" containerName="cinder-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.710238 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" containerName="cinder-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.710476 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" containerName="cinder-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.710522 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="78bf9306-4051-4919-9276-e89b67330f04" containerName="neutron-db-sync" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.720468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.728434 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.731161 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.731440 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2b9z9" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.731561 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.754016 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-sgb9n"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.755632 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.773028 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.783462 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-sgb9n"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802315 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802402 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802425 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802458 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgjj\" (UniqueName: \"kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802498 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.802568 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.808841 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.816372 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.818265 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.818414 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.822176 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5z85h" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.822333 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.828047 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.832134 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-sgb9n"] Oct 11 03:22:25 crc kubenswrapper[4754]: E1011 03:22:25.897942 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-fvdvb ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" podUID="f981caed-948b-43a3-864a-f7313670effe" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903768 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903803 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q8gp\" (UniqueName: \"kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903877 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903899 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903923 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.903952 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.904060 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911341 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911420 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvdvb\" (UniqueName: \"kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911496 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgjj\" (UniqueName: \"kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911563 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911610 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.911774 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.919552 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.923698 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.925563 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.947080 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.957787 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.959509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:25 crc kubenswrapper[4754]: I1011 03:22:25.980610 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgjj\" (UniqueName: \"kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj\") pod \"cinder-scheduler-0\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.001028 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.016019 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.017552 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.018909 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.018943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q8gp\" (UniqueName: \"kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.018979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.018996 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019065 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019106 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvdvb\" (UniqueName: \"kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.019144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.020217 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.033099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.034056 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.040200 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.042848 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.044123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.051825 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.052761 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.053179 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.057449 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q8gp\" (UniqueName: \"kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp\") pod \"neutron-7f8447c5c6-gnvv7\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.068654 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvdvb\" (UniqueName: \"kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.078611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.079552 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc\") pod \"dnsmasq-dns-6bb684768f-sgb9n\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121530 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121737 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121777 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121797 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121826 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121873 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.121952 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpjld\" (UniqueName: \"kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.122035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.122081 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrfvc\" (UniqueName: \"kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.122118 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.122138 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.221675 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230152 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230179 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230283 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpjld\" (UniqueName: \"kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230312 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230372 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrfvc\" (UniqueName: \"kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.230527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.232141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.232522 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.232586 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.235451 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.236459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.237066 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.256036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.262571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.271702 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.272772 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpjld\" (UniqueName: \"kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld\") pod \"dnsmasq-dns-6d97fcdd8f-6dpxh\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.274351 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrfvc\" (UniqueName: \"kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.276289 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.341034 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.379151 4754 generic.go:334] "Generic (PLEG): container finished" podID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerID="19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52" exitCode=0 Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.379241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.379935 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-699df9757c-sqztm" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.381008 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699df9757c-sqztm" event={"ID":"f1663c50-eb8f-4a31-9f7b-f8d720848e56","Type":"ContainerDied","Data":"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52"} Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.381060 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-699df9757c-sqztm" event={"ID":"f1663c50-eb8f-4a31-9f7b-f8d720848e56","Type":"ContainerDied","Data":"30b1f961b5810a0e47a90e3556452871e7cb7afa790bc7d90f6ac9e557050b8f"} Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.381084 4754 scope.go:117] "RemoveContainer" containerID="19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.390577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.400383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.412463 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.433354 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config\") pod \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.433413 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb\") pod \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.433502 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc\") pod \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.433548 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcqbb\" (UniqueName: \"kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb\") pod \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.433597 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb\") pod \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\" (UID: \"f1663c50-eb8f-4a31-9f7b-f8d720848e56\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.438155 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb" (OuterVolumeSpecName: "kube-api-access-hcqbb") pod "f1663c50-eb8f-4a31-9f7b-f8d720848e56" (UID: "f1663c50-eb8f-4a31-9f7b-f8d720848e56"). InnerVolumeSpecName "kube-api-access-hcqbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.504914 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1663c50-eb8f-4a31-9f7b-f8d720848e56" (UID: "f1663c50-eb8f-4a31-9f7b-f8d720848e56"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.505173 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1663c50-eb8f-4a31-9f7b-f8d720848e56" (UID: "f1663c50-eb8f-4a31-9f7b-f8d720848e56"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.517338 4754 scope.go:117] "RemoveContainer" containerID="2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.517660 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1663c50-eb8f-4a31-9f7b-f8d720848e56" (UID: "f1663c50-eb8f-4a31-9f7b-f8d720848e56"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.522430 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config" (OuterVolumeSpecName: "config") pod "f1663c50-eb8f-4a31-9f7b-f8d720848e56" (UID: "f1663c50-eb8f-4a31-9f7b-f8d720848e56"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.543607 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb\") pod \"f981caed-948b-43a3-864a-f7313670effe\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.543653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb\") pod \"f981caed-948b-43a3-864a-f7313670effe\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.543678 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config\") pod \"f981caed-948b-43a3-864a-f7313670effe\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.543737 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc\") pod \"f981caed-948b-43a3-864a-f7313670effe\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.543784 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvdvb\" (UniqueName: \"kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb\") pod \"f981caed-948b-43a3-864a-f7313670effe\" (UID: \"f981caed-948b-43a3-864a-f7313670effe\") " Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544136 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config" (OuterVolumeSpecName: "config") pod "f981caed-948b-43a3-864a-f7313670effe" (UID: "f981caed-948b-43a3-864a-f7313670effe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f981caed-948b-43a3-864a-f7313670effe" (UID: "f981caed-948b-43a3-864a-f7313670effe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544207 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f981caed-948b-43a3-864a-f7313670effe" (UID: "f981caed-948b-43a3-864a-f7313670effe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544558 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544576 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544587 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544595 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544603 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544613 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544621 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcqbb\" (UniqueName: \"kubernetes.io/projected/f1663c50-eb8f-4a31-9f7b-f8d720848e56-kube-api-access-hcqbb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.544630 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1663c50-eb8f-4a31-9f7b-f8d720848e56-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.546226 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f981caed-948b-43a3-864a-f7313670effe" (UID: "f981caed-948b-43a3-864a-f7313670effe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.553088 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb" (OuterVolumeSpecName: "kube-api-access-fvdvb") pod "f981caed-948b-43a3-864a-f7313670effe" (UID: "f981caed-948b-43a3-864a-f7313670effe"). InnerVolumeSpecName "kube-api-access-fvdvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.560951 4754 scope.go:117] "RemoveContainer" containerID="19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52" Oct 11 03:22:26 crc kubenswrapper[4754]: E1011 03:22:26.561729 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52\": container with ID starting with 19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52 not found: ID does not exist" containerID="19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.561767 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52"} err="failed to get container status \"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52\": rpc error: code = NotFound desc = could not find container \"19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52\": container with ID starting with 19c51c1019a7460dffa66fb2fa15ba8b008cb1b581a20e5e1e77b2bfe2527c52 not found: ID does not exist" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.561793 4754 scope.go:117] "RemoveContainer" containerID="2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6" Oct 11 03:22:26 crc kubenswrapper[4754]: E1011 03:22:26.562174 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6\": container with ID starting with 2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6 not found: ID does not exist" containerID="2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.562197 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6"} err="failed to get container status \"2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6\": rpc error: code = NotFound desc = could not find container \"2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6\": container with ID starting with 2ca9651b7e614414a11e9803b9f52fa333600f5fe7e9f020503ac0c2cd705ae6 not found: ID does not exist" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.645764 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f981caed-948b-43a3-864a-f7313670effe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.645797 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvdvb\" (UniqueName: \"kubernetes.io/projected/f981caed-948b-43a3-864a-f7313670effe-kube-api-access-fvdvb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.707745 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.790384 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.798418 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-699df9757c-sqztm"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.934827 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:22:26 crc kubenswrapper[4754]: I1011 03:22:26.966301 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.153306 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" path="/var/lib/kubelet/pods/f1663c50-eb8f-4a31-9f7b-f8d720848e56/volumes" Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.156469 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.260624 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.407141 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerStarted","Data":"8a3eb65d0183855ccb807e60181d148c47f90ff7867ebdebc8bb313c87a57de9"} Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.420061 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" event={"ID":"35a5aafb-45cd-47a0-aae1-e564d0d34848","Type":"ContainerStarted","Data":"a6a32d3cd441a80d701c57c17a2da1cdf99936e412f077af5c5e1ddb7fbf809f"} Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.432266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerStarted","Data":"3d4dbcf1751d22b7872ae4fbd1916a8a5d8767cfa8a3b38fd8b2c8d3b99ecb9b"} Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.434315 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerStarted","Data":"89f56a36297ca5f31a7c145d4a180bf21a8db2fd9d37c8192e29dbcd0438dca3"} Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.434360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-sgb9n" Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.506168 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-sgb9n"] Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.512228 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-sgb9n"] Oct 11 03:22:27 crc kubenswrapper[4754]: I1011 03:22:27.917605 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.469699 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerStarted","Data":"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb"} Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.470143 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.470163 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerStarted","Data":"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635"} Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.472236 4754 generic.go:334] "Generic (PLEG): container finished" podID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerID="272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b" exitCode=0 Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.472285 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" event={"ID":"35a5aafb-45cd-47a0-aae1-e564d0d34848","Type":"ContainerDied","Data":"272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b"} Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.479172 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerStarted","Data":"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520"} Oct 11 03:22:28 crc kubenswrapper[4754]: I1011 03:22:28.507748 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f8447c5c6-gnvv7" podStartSLOduration=3.5077281129999998 podStartE2EDuration="3.507728113s" podCreationTimestamp="2025-10-11 03:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:28.493035508 +0000 UTC m=+996.051980303" watchObservedRunningTime="2025-10-11 03:22:28.507728113 +0000 UTC m=+996.066672898" Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.096506 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f981caed-948b-43a3-864a-f7313670effe" path="/var/lib/kubelet/pods/f981caed-948b-43a3-864a-f7313670effe/volumes" Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.489003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerStarted","Data":"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8"} Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.489046 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerStarted","Data":"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da"} Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.494655 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" event={"ID":"35a5aafb-45cd-47a0-aae1-e564d0d34848","Type":"ContainerStarted","Data":"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524"} Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.495551 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.500444 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerStarted","Data":"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7"} Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.500617 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api-log" containerID="cri-o://e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" gracePeriod=30 Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.500657 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api" containerID="cri-o://52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" gracePeriod=30 Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.532272 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.567987092 podStartE2EDuration="4.532255279s" podCreationTimestamp="2025-10-11 03:22:25 +0000 UTC" firstStartedPulling="2025-10-11 03:22:26.718129131 +0000 UTC m=+994.277073916" lastFinishedPulling="2025-10-11 03:22:27.682397318 +0000 UTC m=+995.241342103" observedRunningTime="2025-10-11 03:22:29.518317516 +0000 UTC m=+997.077262311" watchObservedRunningTime="2025-10-11 03:22:29.532255279 +0000 UTC m=+997.091200064" Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.566487 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" podStartSLOduration=4.566467575 podStartE2EDuration="4.566467575s" podCreationTimestamp="2025-10-11 03:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:29.552734997 +0000 UTC m=+997.111679772" watchObservedRunningTime="2025-10-11 03:22:29.566467575 +0000 UTC m=+997.125412360" Oct 11 03:22:29 crc kubenswrapper[4754]: I1011 03:22:29.584472 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.584444472 podStartE2EDuration="4.584444472s" podCreationTimestamp="2025-10-11 03:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:29.573539035 +0000 UTC m=+997.132483840" watchObservedRunningTime="2025-10-11 03:22:29.584444472 +0000 UTC m=+997.143389257" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.361812 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468514 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468580 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrfvc\" (UniqueName: \"kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468672 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468705 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468747 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468798 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.468846 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id\") pod \"1111ff77-0127-41d1-b780-791de5b78267\" (UID: \"1111ff77-0127-41d1-b780-791de5b78267\") " Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.469266 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.469617 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs" (OuterVolumeSpecName: "logs") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.474071 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc" (OuterVolumeSpecName: "kube-api-access-wrfvc") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "kube-api-access-wrfvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.475121 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts" (OuterVolumeSpecName: "scripts") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.481069 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.504867 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.522919 4754 generic.go:334] "Generic (PLEG): container finished" podID="1111ff77-0127-41d1-b780-791de5b78267" containerID="52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" exitCode=0 Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.522948 4754 generic.go:334] "Generic (PLEG): container finished" podID="1111ff77-0127-41d1-b780-791de5b78267" containerID="e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" exitCode=143 Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.523850 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.524290 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerDied","Data":"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7"} Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.524314 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerDied","Data":"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520"} Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.524324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1111ff77-0127-41d1-b780-791de5b78267","Type":"ContainerDied","Data":"3d4dbcf1751d22b7872ae4fbd1916a8a5d8767cfa8a3b38fd8b2c8d3b99ecb9b"} Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.524338 4754 scope.go:117] "RemoveContainer" containerID="52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.537454 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data" (OuterVolumeSpecName: "config-data") pod "1111ff77-0127-41d1-b780-791de5b78267" (UID: "1111ff77-0127-41d1-b780-791de5b78267"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570666 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1111ff77-0127-41d1-b780-791de5b78267-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570698 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1111ff77-0127-41d1-b780-791de5b78267-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570709 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrfvc\" (UniqueName: \"kubernetes.io/projected/1111ff77-0127-41d1-b780-791de5b78267-kube-api-access-wrfvc\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570718 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570728 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570737 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.570745 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1111ff77-0127-41d1-b780-791de5b78267-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.583922 4754 scope.go:117] "RemoveContainer" containerID="e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.600401 4754 scope.go:117] "RemoveContainer" containerID="52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.600891 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7\": container with ID starting with 52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7 not found: ID does not exist" containerID="52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601023 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7"} err="failed to get container status \"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7\": rpc error: code = NotFound desc = could not find container \"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7\": container with ID starting with 52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7 not found: ID does not exist" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601132 4754 scope.go:117] "RemoveContainer" containerID="e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.601521 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520\": container with ID starting with e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520 not found: ID does not exist" containerID="e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601552 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520"} err="failed to get container status \"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520\": rpc error: code = NotFound desc = could not find container \"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520\": container with ID starting with e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520 not found: ID does not exist" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601575 4754 scope.go:117] "RemoveContainer" containerID="52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601886 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7"} err="failed to get container status \"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7\": rpc error: code = NotFound desc = could not find container \"52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7\": container with ID starting with 52a72b26aa3733168e3896fe106f265b5c3875266189f62f49453f2e956a97c7 not found: ID does not exist" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.601935 4754 scope.go:117] "RemoveContainer" containerID="e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.602241 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520"} err="failed to get container status \"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520\": rpc error: code = NotFound desc = could not find container \"e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520\": container with ID starting with e44c94cdeffb40f06e36fd7bf3eb0fb9729a710b7b0048d7091ace4110432520 not found: ID does not exist" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.736767 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.736845 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.870958 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.880800 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.913421 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.913846 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api-log" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.913862 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api-log" Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.913875 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.913882 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api" Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.913911 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="init" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.913917 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="init" Oct 11 03:22:30 crc kubenswrapper[4754]: E1011 03:22:30.913928 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="dnsmasq-dns" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.913935 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="dnsmasq-dns" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.914120 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.914139 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1663c50-eb8f-4a31-9f7b-f8d720848e56" containerName="dnsmasq-dns" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.914149 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1111ff77-0127-41d1-b780-791de5b78267" containerName="cinder-api-log" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.915156 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.921409 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.921825 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.921947 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 11 03:22:30 crc kubenswrapper[4754]: I1011 03:22:30.926213 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.054228 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079343 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079425 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079472 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbn6z\" (UniqueName: \"kubernetes.io/projected/be981514-5d9c-4c0a-8853-12d2d04e6e76-kube-api-access-sbn6z\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079623 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-scripts\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.079858 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be981514-5d9c-4c0a-8853-12d2d04e6e76-logs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.080135 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.080252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be981514-5d9c-4c0a-8853-12d2d04e6e76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.080286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data-custom\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.093687 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1111ff77-0127-41d1-b780-791de5b78267" path="/var/lib/kubelet/pods/1111ff77-0127-41d1-b780-791de5b78267/volumes" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183430 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbn6z\" (UniqueName: \"kubernetes.io/projected/be981514-5d9c-4c0a-8853-12d2d04e6e76-kube-api-access-sbn6z\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183477 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-scripts\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183533 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be981514-5d9c-4c0a-8853-12d2d04e6e76-logs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183554 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data-custom\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183607 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be981514-5d9c-4c0a-8853-12d2d04e6e76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183658 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183678 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.183705 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.184624 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be981514-5d9c-4c0a-8853-12d2d04e6e76-logs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.187103 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be981514-5d9c-4c0a-8853-12d2d04e6e76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.188803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-scripts\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.189468 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.189926 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.192569 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data-custom\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.195507 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.207479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbn6z\" (UniqueName: \"kubernetes.io/projected/be981514-5d9c-4c0a-8853-12d2d04e6e76-kube-api-access-sbn6z\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.211220 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be981514-5d9c-4c0a-8853-12d2d04e6e76-config-data\") pod \"cinder-api-0\" (UID: \"be981514-5d9c-4c0a-8853-12d2d04e6e76\") " pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.231996 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.810509 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 03:22:31 crc kubenswrapper[4754]: I1011 03:22:31.993483 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-57cbfbbbc9-kgfv2"] Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.009372 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.012310 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.012517 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.020520 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57cbfbbbc9-kgfv2"] Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.109655 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-internal-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.109709 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmls6\" (UniqueName: \"kubernetes.io/projected/50322a6c-d0d3-4574-a584-e94ef7b39e70-kube-api-access-qmls6\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.109874 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-combined-ca-bundle\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.109935 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.110058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-httpd-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.110180 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-public-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.110209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-ovndb-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.210780 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-httpd-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.210827 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-public-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.210849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-ovndb-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.210883 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-internal-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.211727 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmls6\" (UniqueName: \"kubernetes.io/projected/50322a6c-d0d3-4574-a584-e94ef7b39e70-kube-api-access-qmls6\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.212127 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-combined-ca-bundle\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.212165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.217269 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-httpd-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.218218 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-combined-ca-bundle\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.218691 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-internal-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.219739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-ovndb-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.225992 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-public-tls-certs\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.228393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmls6\" (UniqueName: \"kubernetes.io/projected/50322a6c-d0d3-4574-a584-e94ef7b39e70-kube-api-access-qmls6\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.228796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/50322a6c-d0d3-4574-a584-e94ef7b39e70-config\") pod \"neutron-57cbfbbbc9-kgfv2\" (UID: \"50322a6c-d0d3-4574-a584-e94ef7b39e70\") " pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.335551 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.549086 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be981514-5d9c-4c0a-8853-12d2d04e6e76","Type":"ContainerStarted","Data":"b2001b0fb7e8ccb23a1cc77efaa629a8eea6ba03347f1382aeba7615fc0ecb32"} Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.549124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be981514-5d9c-4c0a-8853-12d2d04e6e76","Type":"ContainerStarted","Data":"1b834c94ae375ecf51f7df1fd02cddf5cc7fe8b6a35135e0d460c6c2736e3569"} Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.575996 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.682663 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6658bb6dd-v68dl" Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.752685 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.753100 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85464d6c8b-t4kkc" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api" containerID="cri-o://d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b" gracePeriod=30 Oct 11 03:22:32 crc kubenswrapper[4754]: I1011 03:22:32.752944 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85464d6c8b-t4kkc" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api-log" containerID="cri-o://9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629" gracePeriod=30 Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.177790 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57cbfbbbc9-kgfv2"] Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.562018 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be981514-5d9c-4c0a-8853-12d2d04e6e76","Type":"ContainerStarted","Data":"d2e37f9440e29a621a570cc58bbc8226af80a502157c87d28787bc4419cb5850"} Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.562369 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.566216 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cbfbbbc9-kgfv2" event={"ID":"50322a6c-d0d3-4574-a584-e94ef7b39e70","Type":"ContainerStarted","Data":"ebbde006a5f5a47fb0e9544215897aeb243ed2ec328daad5bcc179da19c41f0f"} Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.566359 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cbfbbbc9-kgfv2" event={"ID":"50322a6c-d0d3-4574-a584-e94ef7b39e70","Type":"ContainerStarted","Data":"9c63c38c0e69798487533989413dfcd4e4744c268936bf134591c2f3e4400d43"} Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.568775 4754 generic.go:334] "Generic (PLEG): container finished" podID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerID="9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629" exitCode=143 Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.568831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerDied","Data":"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629"} Oct 11 03:22:33 crc kubenswrapper[4754]: I1011 03:22:33.592014 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.591990706 podStartE2EDuration="3.591990706s" podCreationTimestamp="2025-10-11 03:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:33.585647377 +0000 UTC m=+1001.144592162" watchObservedRunningTime="2025-10-11 03:22:33.591990706 +0000 UTC m=+1001.150935501" Oct 11 03:22:34 crc kubenswrapper[4754]: I1011 03:22:34.582538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cbfbbbc9-kgfv2" event={"ID":"50322a6c-d0d3-4574-a584-e94ef7b39e70","Type":"ContainerStarted","Data":"4c099ed5bc43c2ad463ea31e2da154ff9f9043d381c9538b258cd9d47c9b2fb0"} Oct 11 03:22:34 crc kubenswrapper[4754]: I1011 03:22:34.582941 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:22:34 crc kubenswrapper[4754]: I1011 03:22:34.609332 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-57cbfbbbc9-kgfv2" podStartSLOduration=3.60931291 podStartE2EDuration="3.60931291s" podCreationTimestamp="2025-10-11 03:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:34.603676931 +0000 UTC m=+1002.162621716" watchObservedRunningTime="2025-10-11 03:22:34.60931291 +0000 UTC m=+1002.168257695" Oct 11 03:22:35 crc kubenswrapper[4754]: I1011 03:22:35.789022 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-748c895d7d-pnbk5" Oct 11 03:22:35 crc kubenswrapper[4754]: I1011 03:22:35.953334 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85464d6c8b-t4kkc" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:59014->10.217.0.146:9311: read: connection reset by peer" Oct 11 03:22:35 crc kubenswrapper[4754]: I1011 03:22:35.953794 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85464d6c8b-t4kkc" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:58998->10.217.0.146:9311: read: connection reset by peer" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.358007 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.393078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.411585 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.480046 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.496342 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.496585 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="dnsmasq-dns" containerID="cri-o://687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1" gracePeriod=10 Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.515720 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w78ll\" (UniqueName: \"kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll\") pod \"7b8d2473-13f3-4c7f-9347-0476f630f042\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.515833 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle\") pod \"7b8d2473-13f3-4c7f-9347-0476f630f042\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.515859 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom\") pod \"7b8d2473-13f3-4c7f-9347-0476f630f042\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.515932 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data\") pod \"7b8d2473-13f3-4c7f-9347-0476f630f042\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.515989 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs\") pod \"7b8d2473-13f3-4c7f-9347-0476f630f042\" (UID: \"7b8d2473-13f3-4c7f-9347-0476f630f042\") " Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.516814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs" (OuterVolumeSpecName: "logs") pod "7b8d2473-13f3-4c7f-9347-0476f630f042" (UID: "7b8d2473-13f3-4c7f-9347-0476f630f042"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.535495 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll" (OuterVolumeSpecName: "kube-api-access-w78ll") pod "7b8d2473-13f3-4c7f-9347-0476f630f042" (UID: "7b8d2473-13f3-4c7f-9347-0476f630f042"). InnerVolumeSpecName "kube-api-access-w78ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.553117 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7b8d2473-13f3-4c7f-9347-0476f630f042" (UID: "7b8d2473-13f3-4c7f-9347-0476f630f042"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.599525 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b8d2473-13f3-4c7f-9347-0476f630f042" (UID: "7b8d2473-13f3-4c7f-9347-0476f630f042"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.618188 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8d2473-13f3-4c7f-9347-0476f630f042-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.618225 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w78ll\" (UniqueName: \"kubernetes.io/projected/7b8d2473-13f3-4c7f-9347-0476f630f042-kube-api-access-w78ll\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.618240 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.618253 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.629336 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data" (OuterVolumeSpecName: "config-data") pod "7b8d2473-13f3-4c7f-9347-0476f630f042" (UID: "7b8d2473-13f3-4c7f-9347-0476f630f042"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.650672 4754 generic.go:334] "Generic (PLEG): container finished" podID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerID="d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b" exitCode=0 Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.651010 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="cinder-scheduler" containerID="cri-o://095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da" gracePeriod=30 Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.651191 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85464d6c8b-t4kkc" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.657138 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerDied","Data":"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b"} Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.657226 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85464d6c8b-t4kkc" event={"ID":"7b8d2473-13f3-4c7f-9347-0476f630f042","Type":"ContainerDied","Data":"6462fed4fd9fd5d93dfd4ce21e3020f324fcab4d617532200c0600ce02ed716f"} Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.657259 4754 scope.go:117] "RemoveContainer" containerID="d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.658367 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="probe" containerID="cri-o://ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8" gracePeriod=30 Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.717326 4754 scope.go:117] "RemoveContainer" containerID="9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.719672 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8d2473-13f3-4c7f-9347-0476f630f042-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.724381 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.732664 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-85464d6c8b-t4kkc"] Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.745690 4754 scope.go:117] "RemoveContainer" containerID="d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b" Oct 11 03:22:36 crc kubenswrapper[4754]: E1011 03:22:36.750754 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b\": container with ID starting with d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b not found: ID does not exist" containerID="d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.750801 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b"} err="failed to get container status \"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b\": rpc error: code = NotFound desc = could not find container \"d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b\": container with ID starting with d9e5cc6f3c76780058abae036b9762bb9e38b197d685cc7b70e6b5490fdb393b not found: ID does not exist" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.750830 4754 scope.go:117] "RemoveContainer" containerID="9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629" Oct 11 03:22:36 crc kubenswrapper[4754]: E1011 03:22:36.751525 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629\": container with ID starting with 9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629 not found: ID does not exist" containerID="9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629" Oct 11 03:22:36 crc kubenswrapper[4754]: I1011 03:22:36.751571 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629"} err="failed to get container status \"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629\": rpc error: code = NotFound desc = could not find container \"9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629\": container with ID starting with 9bda97257f08db742bcb85b12e78e4ed84fdeda992700c33fae72aa3db6d3629 not found: ID does not exist" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.094448 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" path="/var/lib/kubelet/pods/7b8d2473-13f3-4c7f-9347-0476f630f042/volumes" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.098218 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.232400 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb\") pod \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.233072 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2z97\" (UniqueName: \"kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97\") pod \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.233138 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc\") pod \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.233208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb\") pod \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.233298 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config\") pod \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\" (UID: \"c3446bc0-2a40-4ead-926d-28ee8f8f04f0\") " Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.242007 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97" (OuterVolumeSpecName: "kube-api-access-x2z97") pod "c3446bc0-2a40-4ead-926d-28ee8f8f04f0" (UID: "c3446bc0-2a40-4ead-926d-28ee8f8f04f0"). InnerVolumeSpecName "kube-api-access-x2z97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.283991 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config" (OuterVolumeSpecName: "config") pod "c3446bc0-2a40-4ead-926d-28ee8f8f04f0" (UID: "c3446bc0-2a40-4ead-926d-28ee8f8f04f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.284131 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3446bc0-2a40-4ead-926d-28ee8f8f04f0" (UID: "c3446bc0-2a40-4ead-926d-28ee8f8f04f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.310504 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3446bc0-2a40-4ead-926d-28ee8f8f04f0" (UID: "c3446bc0-2a40-4ead-926d-28ee8f8f04f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.320921 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3446bc0-2a40-4ead-926d-28ee8f8f04f0" (UID: "c3446bc0-2a40-4ead-926d-28ee8f8f04f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.336145 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.336184 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.336200 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2z97\" (UniqueName: \"kubernetes.io/projected/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-kube-api-access-x2z97\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.336210 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.336220 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3446bc0-2a40-4ead-926d-28ee8f8f04f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.340787 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.341255 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341270 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api" Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.341294 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="dnsmasq-dns" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341301 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="dnsmasq-dns" Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.341318 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="init" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341325 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="init" Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.341338 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api-log" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341344 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api-log" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341524 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341540 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerName="dnsmasq-dns" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.341553 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8d2473-13f3-4c7f-9347-0476f630f042" containerName="barbican-api-log" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.342169 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.344799 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.345038 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-fdgps" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.345384 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.350860 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.438066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.438116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25s8\" (UniqueName: \"kubernetes.io/projected/89166926-b13c-442e-ba5f-b15200cc237a-kube-api-access-l25s8\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.438328 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config-secret\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.438398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.540112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config-secret\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.540390 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.540608 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.540682 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25s8\" (UniqueName: \"kubernetes.io/projected/89166926-b13c-442e-ba5f-b15200cc237a-kube-api-access-l25s8\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.541518 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.544884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-openstack-config-secret\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.545879 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89166926-b13c-442e-ba5f-b15200cc237a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.563344 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25s8\" (UniqueName: \"kubernetes.io/projected/89166926-b13c-442e-ba5f-b15200cc237a-kube-api-access-l25s8\") pod \"openstackclient\" (UID: \"89166926-b13c-442e-ba5f-b15200cc237a\") " pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.659802 4754 generic.go:334] "Generic (PLEG): container finished" podID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerID="ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8" exitCode=0 Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.659855 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerDied","Data":"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8"} Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.661836 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" containerID="687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1" exitCode=0 Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.661900 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" event={"ID":"c3446bc0-2a40-4ead-926d-28ee8f8f04f0","Type":"ContainerDied","Data":"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1"} Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.661930 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" event={"ID":"c3446bc0-2a40-4ead-926d-28ee8f8f04f0","Type":"ContainerDied","Data":"db9e836a997290e94c5c3a60fc797f8d996adb9200832c2460d5017e875b8630"} Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.661980 4754 scope.go:117] "RemoveContainer" containerID="687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.661978 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-mqmbt" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.666000 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.708794 4754 scope.go:117] "RemoveContainer" containerID="f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.709572 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.723619 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-mqmbt"] Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.729957 4754 scope.go:117] "RemoveContainer" containerID="687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1" Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.731526 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1\": container with ID starting with 687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1 not found: ID does not exist" containerID="687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.731562 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1"} err="failed to get container status \"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1\": rpc error: code = NotFound desc = could not find container \"687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1\": container with ID starting with 687592d50e89f596e37b4d26927713a5a88e22b62bd9a6548a84b5b99b3e6ec1 not found: ID does not exist" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.731609 4754 scope.go:117] "RemoveContainer" containerID="f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039" Oct 11 03:22:37 crc kubenswrapper[4754]: E1011 03:22:37.733199 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039\": container with ID starting with f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039 not found: ID does not exist" containerID="f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039" Oct 11 03:22:37 crc kubenswrapper[4754]: I1011 03:22:37.733248 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039"} err="failed to get container status \"f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039\": rpc error: code = NotFound desc = could not find container \"f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039\": container with ID starting with f68f5209da479dd6eba05a6dd9ef40e58f4ad96a50a0f0a080950688a169a039 not found: ID does not exist" Oct 11 03:22:38 crc kubenswrapper[4754]: I1011 03:22:38.144830 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 11 03:22:38 crc kubenswrapper[4754]: W1011 03:22:38.157774 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89166926_b13c_442e_ba5f_b15200cc237a.slice/crio-89d705ade8b5aacf8f8e3cb3afcd60de90fa6e2a11b8ed2f7c972d79bc6115f1 WatchSource:0}: Error finding container 89d705ade8b5aacf8f8e3cb3afcd60de90fa6e2a11b8ed2f7c972d79bc6115f1: Status 404 returned error can't find the container with id 89d705ade8b5aacf8f8e3cb3afcd60de90fa6e2a11b8ed2f7c972d79bc6115f1 Oct 11 03:22:38 crc kubenswrapper[4754]: I1011 03:22:38.681771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"89166926-b13c-442e-ba5f-b15200cc237a","Type":"ContainerStarted","Data":"89d705ade8b5aacf8f8e3cb3afcd60de90fa6e2a11b8ed2f7c972d79bc6115f1"} Oct 11 03:22:39 crc kubenswrapper[4754]: I1011 03:22:39.096541 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3446bc0-2a40-4ead-926d-28ee8f8f04f0" path="/var/lib/kubelet/pods/c3446bc0-2a40-4ead-926d-28ee8f8f04f0/volumes" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.273704 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.405732 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.405780 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.405820 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.405979 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.406041 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.406767 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdgjj\" (UniqueName: \"kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.406805 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts\") pod \"48218ef7-c454-4c3d-bbbc-6b838b903c29\" (UID: \"48218ef7-c454-4c3d-bbbc-6b838b903c29\") " Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.407683 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48218ef7-c454-4c3d-bbbc-6b838b903c29-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.412844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj" (OuterVolumeSpecName: "kube-api-access-fdgjj") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "kube-api-access-fdgjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.412928 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.413513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts" (OuterVolumeSpecName: "scripts") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.470429 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.509121 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdgjj\" (UniqueName: \"kubernetes.io/projected/48218ef7-c454-4c3d-bbbc-6b838b903c29-kube-api-access-fdgjj\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.509162 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.509174 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.509184 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.524305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data" (OuterVolumeSpecName: "config-data") pod "48218ef7-c454-4c3d-bbbc-6b838b903c29" (UID: "48218ef7-c454-4c3d-bbbc-6b838b903c29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.611494 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48218ef7-c454-4c3d-bbbc-6b838b903c29-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.703794 4754 generic.go:334] "Generic (PLEG): container finished" podID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerID="095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da" exitCode=0 Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.703837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerDied","Data":"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da"} Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.703864 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48218ef7-c454-4c3d-bbbc-6b838b903c29","Type":"ContainerDied","Data":"89f56a36297ca5f31a7c145d4a180bf21a8db2fd9d37c8192e29dbcd0438dca3"} Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.703866 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.703881 4754 scope.go:117] "RemoveContainer" containerID="ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.740214 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.759017 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.761261 4754 scope.go:117] "RemoveContainer" containerID="095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.776125 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:40 crc kubenswrapper[4754]: E1011 03:22:40.776488 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="cinder-scheduler" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.776499 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="cinder-scheduler" Oct 11 03:22:40 crc kubenswrapper[4754]: E1011 03:22:40.776524 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="probe" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.776530 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="probe" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.776683 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="cinder-scheduler" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.776693 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" containerName="probe" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.777600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.779804 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.792935 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.805712 4754 scope.go:117] "RemoveContainer" containerID="ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8" Oct 11 03:22:40 crc kubenswrapper[4754]: E1011 03:22:40.806341 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8\": container with ID starting with ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8 not found: ID does not exist" containerID="ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.806369 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8"} err="failed to get container status \"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8\": rpc error: code = NotFound desc = could not find container \"ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8\": container with ID starting with ab7e6be86900a27aaa296d6aaa60a4a597d16142bd0b9037766d2317661852a8 not found: ID does not exist" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.806389 4754 scope.go:117] "RemoveContainer" containerID="095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da" Oct 11 03:22:40 crc kubenswrapper[4754]: E1011 03:22:40.808182 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da\": container with ID starting with 095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da not found: ID does not exist" containerID="095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.808206 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da"} err="failed to get container status \"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da\": rpc error: code = NotFound desc = could not find container \"095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da\": container with ID starting with 095d17c10f09534cd2c4211f502c8410115f5b73b44b8aa37a8ba0bff4b946da not found: ID does not exist" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.813884 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.813922 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.814007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzbkm\" (UniqueName: \"kubernetes.io/projected/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-kube-api-access-zzbkm\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.814029 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.814050 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.814069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-scripts\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.915835 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzbkm\" (UniqueName: \"kubernetes.io/projected/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-kube-api-access-zzbkm\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.916232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.916310 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.916346 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.916382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-scripts\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.917226 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.917272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.920261 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-scripts\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.920719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.920883 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.922167 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-config-data\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:40 crc kubenswrapper[4754]: I1011 03:22:40.938798 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzbkm\" (UniqueName: \"kubernetes.io/projected/5543cbf6-34ea-4bdc-86c3-7b61d6a14c03-kube-api-access-zzbkm\") pod \"cinder-scheduler-0\" (UID: \"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03\") " pod="openstack/cinder-scheduler-0" Oct 11 03:22:41 crc kubenswrapper[4754]: I1011 03:22:41.104481 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 03:22:41 crc kubenswrapper[4754]: I1011 03:22:41.107754 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48218ef7-c454-4c3d-bbbc-6b838b903c29" path="/var/lib/kubelet/pods/48218ef7-c454-4c3d-bbbc-6b838b903c29/volumes" Oct 11 03:22:41 crc kubenswrapper[4754]: I1011 03:22:41.590871 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 03:22:41 crc kubenswrapper[4754]: W1011 03:22:41.599389 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5543cbf6_34ea_4bdc_86c3_7b61d6a14c03.slice/crio-379c304617dd6d84200b2502fdd111f212ca144a81af1aacf62d3f4205c4994b WatchSource:0}: Error finding container 379c304617dd6d84200b2502fdd111f212ca144a81af1aacf62d3f4205c4994b: Status 404 returned error can't find the container with id 379c304617dd6d84200b2502fdd111f212ca144a81af1aacf62d3f4205c4994b Oct 11 03:22:41 crc kubenswrapper[4754]: I1011 03:22:41.720189 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03","Type":"ContainerStarted","Data":"379c304617dd6d84200b2502fdd111f212ca144a81af1aacf62d3f4205c4994b"} Oct 11 03:22:42 crc kubenswrapper[4754]: I1011 03:22:42.737802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03","Type":"ContainerStarted","Data":"15a7e5af626cc54bd703bd91dc53c744f4132f59e61da392a5cc2a32d2334598"} Oct 11 03:22:43 crc kubenswrapper[4754]: I1011 03:22:43.284830 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 11 03:22:43 crc kubenswrapper[4754]: I1011 03:22:43.531476 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 03:22:43 crc kubenswrapper[4754]: I1011 03:22:43.759912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5543cbf6-34ea-4bdc-86c3-7b61d6a14c03","Type":"ContainerStarted","Data":"f71ee7a9243f0dd5bd0507489b94fd4ff19cc08d18668b2a6860db786f191852"} Oct 11 03:22:43 crc kubenswrapper[4754]: I1011 03:22:43.778420 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.778405218 podStartE2EDuration="3.778405218s" podCreationTimestamp="2025-10-11 03:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:43.775562038 +0000 UTC m=+1011.334506823" watchObservedRunningTime="2025-10-11 03:22:43.778405218 +0000 UTC m=+1011.337350003" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.035596 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.036376 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="976f8dd9-0b51-4b45-9192-9ddf098acd48" containerName="kube-state-metrics" containerID="cri-o://e69b9f7f9abcf2f5e6a11602a13c9855ac7c757ab742a44cec0d6ecfa3113d38" gracePeriod=30 Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.104863 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.779294 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mkp24"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.780565 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.793930 4754 generic.go:334] "Generic (PLEG): container finished" podID="976f8dd9-0b51-4b45-9192-9ddf098acd48" containerID="e69b9f7f9abcf2f5e6a11602a13c9855ac7c757ab742a44cec0d6ecfa3113d38" exitCode=2 Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.793983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"976f8dd9-0b51-4b45-9192-9ddf098acd48","Type":"ContainerDied","Data":"e69b9f7f9abcf2f5e6a11602a13c9855ac7c757ab742a44cec0d6ecfa3113d38"} Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.799978 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mkp24"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.857609 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-j2hk9"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.858994 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.865181 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j2hk9"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.939745 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rxkh\" (UniqueName: \"kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh\") pod \"nova-api-db-create-mkp24\" (UID: \"8f2d4813-8e67-4e79-8aad-5519f6355ced\") " pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.967053 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rwhqn"] Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.969715 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:46 crc kubenswrapper[4754]: I1011 03:22:46.976540 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rwhqn"] Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.041508 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rxkh\" (UniqueName: \"kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh\") pod \"nova-api-db-create-mkp24\" (UID: \"8f2d4813-8e67-4e79-8aad-5519f6355ced\") " pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.041634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knshc\" (UniqueName: \"kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc\") pod \"nova-cell0-db-create-j2hk9\" (UID: \"3a5329e2-6d7f-477c-9496-347d8a49f4df\") " pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.063544 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rxkh\" (UniqueName: \"kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh\") pod \"nova-api-db-create-mkp24\" (UID: \"8f2d4813-8e67-4e79-8aad-5519f6355ced\") " pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.098999 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.100679 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.101005 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-central-agent" containerID="cri-o://d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c" gracePeriod=30 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.101290 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="proxy-httpd" containerID="cri-o://0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4" gracePeriod=30 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.101336 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="sg-core" containerID="cri-o://4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b" gracePeriod=30 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.101372 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-notification-agent" containerID="cri-o://8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f" gracePeriod=30 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.143386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m5kz\" (UniqueName: \"kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz\") pod \"nova-cell1-db-create-rwhqn\" (UID: \"b3625e97-9933-4bf5-9253-48a0465301fd\") " pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.143511 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knshc\" (UniqueName: \"kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc\") pod \"nova-cell0-db-create-j2hk9\" (UID: \"3a5329e2-6d7f-477c-9496-347d8a49f4df\") " pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.161365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knshc\" (UniqueName: \"kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc\") pod \"nova-cell0-db-create-j2hk9\" (UID: \"3a5329e2-6d7f-477c-9496-347d8a49f4df\") " pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.175371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.245827 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m5kz\" (UniqueName: \"kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz\") pod \"nova-cell1-db-create-rwhqn\" (UID: \"b3625e97-9933-4bf5-9253-48a0465301fd\") " pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.266245 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m5kz\" (UniqueName: \"kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz\") pod \"nova-cell1-db-create-rwhqn\" (UID: \"b3625e97-9933-4bf5-9253-48a0465301fd\") " pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.292382 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807559 4754 generic.go:334] "Generic (PLEG): container finished" podID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerID="0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4" exitCode=0 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807887 4754 generic.go:334] "Generic (PLEG): container finished" podID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerID="4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b" exitCode=2 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807899 4754 generic.go:334] "Generic (PLEG): container finished" podID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerID="d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c" exitCode=0 Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807626 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerDied","Data":"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4"} Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807941 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerDied","Data":"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b"} Oct 11 03:22:47 crc kubenswrapper[4754]: I1011 03:22:47.807979 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerDied","Data":"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c"} Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.482792 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.587027 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wxml\" (UniqueName: \"kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml\") pod \"976f8dd9-0b51-4b45-9192-9ddf098acd48\" (UID: \"976f8dd9-0b51-4b45-9192-9ddf098acd48\") " Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.593118 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml" (OuterVolumeSpecName: "kube-api-access-8wxml") pod "976f8dd9-0b51-4b45-9192-9ddf098acd48" (UID: "976f8dd9-0b51-4b45-9192-9ddf098acd48"). InnerVolumeSpecName "kube-api-access-8wxml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.689659 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wxml\" (UniqueName: \"kubernetes.io/projected/976f8dd9-0b51-4b45-9192-9ddf098acd48-kube-api-access-8wxml\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.736380 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rwhqn"] Oct 11 03:22:49 crc kubenswrapper[4754]: W1011 03:22:49.744788 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3625e97_9933_4bf5_9253_48a0465301fd.slice/crio-8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc WatchSource:0}: Error finding container 8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc: Status 404 returned error can't find the container with id 8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.816915 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mkp24"] Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.831044 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"976f8dd9-0b51-4b45-9192-9ddf098acd48","Type":"ContainerDied","Data":"62ad13f13b8c623dee21eda3bc85c1423d8473325e34037b84a376934792841b"} Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.831094 4754 scope.go:117] "RemoveContainer" containerID="e69b9f7f9abcf2f5e6a11602a13c9855ac7c757ab742a44cec0d6ecfa3113d38" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.831093 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.834608 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rwhqn" event={"ID":"b3625e97-9933-4bf5-9253-48a0465301fd","Type":"ContainerStarted","Data":"8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc"} Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.836124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"89166926-b13c-442e-ba5f-b15200cc237a","Type":"ContainerStarted","Data":"6c8ee86e8dddea817ed1cb326a34a3f135192908f8115eea81c97d64a38f4b0d"} Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.838321 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkp24" event={"ID":"8f2d4813-8e67-4e79-8aad-5519f6355ced","Type":"ContainerStarted","Data":"67b1f40eb624ef880542f556874fe73cf20957a8fa2db23314cbdc5a7cdabcc6"} Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.856566 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.680532381 podStartE2EDuration="12.856534444s" podCreationTimestamp="2025-10-11 03:22:37 +0000 UTC" firstStartedPulling="2025-10-11 03:22:38.160165823 +0000 UTC m=+1005.719110608" lastFinishedPulling="2025-10-11 03:22:49.336167886 +0000 UTC m=+1016.895112671" observedRunningTime="2025-10-11 03:22:49.849537677 +0000 UTC m=+1017.408482462" watchObservedRunningTime="2025-10-11 03:22:49.856534444 +0000 UTC m=+1017.415479229" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.880310 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j2hk9"] Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.890070 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.903099 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.918133 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:49 crc kubenswrapper[4754]: E1011 03:22:49.918559 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="976f8dd9-0b51-4b45-9192-9ddf098acd48" containerName="kube-state-metrics" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.918575 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="976f8dd9-0b51-4b45-9192-9ddf098acd48" containerName="kube-state-metrics" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.918750 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="976f8dd9-0b51-4b45-9192-9ddf098acd48" containerName="kube-state-metrics" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.919367 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.922725 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.924138 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 11 03:22:49 crc kubenswrapper[4754]: I1011 03:22:49.940061 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.099809 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4drt\" (UniqueName: \"kubernetes.io/projected/955e7c85-4242-453b-b23d-432057450e28-kube-api-access-v4drt\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.099874 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.099915 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.099993 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.201950 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4drt\" (UniqueName: \"kubernetes.io/projected/955e7c85-4242-453b-b23d-432057450e28-kube-api-access-v4drt\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.202057 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.202096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.202144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.207167 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.209410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.212490 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/955e7c85-4242-453b-b23d-432057450e28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.219077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4drt\" (UniqueName: \"kubernetes.io/projected/955e7c85-4242-453b-b23d-432057450e28-kube-api-access-v4drt\") pod \"kube-state-metrics-0\" (UID: \"955e7c85-4242-453b-b23d-432057450e28\") " pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.264342 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.727632 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.851634 4754 generic.go:334] "Generic (PLEG): container finished" podID="b3625e97-9933-4bf5-9253-48a0465301fd" containerID="e5ff870bad281ae0edbce039f90a1579be77878a17536f1a8f6d5d254ae2f0db" exitCode=0 Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.851702 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rwhqn" event={"ID":"b3625e97-9933-4bf5-9253-48a0465301fd","Type":"ContainerDied","Data":"e5ff870bad281ae0edbce039f90a1579be77878a17536f1a8f6d5d254ae2f0db"} Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.854130 4754 generic.go:334] "Generic (PLEG): container finished" podID="3a5329e2-6d7f-477c-9496-347d8a49f4df" containerID="a4f0973a5edbcd3d0e55781498362458aed82b78c14fbc2fb36987ee44beb8dc" exitCode=0 Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.854279 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j2hk9" event={"ID":"3a5329e2-6d7f-477c-9496-347d8a49f4df","Type":"ContainerDied","Data":"a4f0973a5edbcd3d0e55781498362458aed82b78c14fbc2fb36987ee44beb8dc"} Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.854308 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j2hk9" event={"ID":"3a5329e2-6d7f-477c-9496-347d8a49f4df","Type":"ContainerStarted","Data":"47f8d68859afd54bc172c42ed74f8fe0c442710cd17909c6f2436ded02985b23"} Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.856490 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkp24" event={"ID":"8f2d4813-8e67-4e79-8aad-5519f6355ced","Type":"ContainerDied","Data":"129073840aa223733a780a4be753edfc279bc404440474759f083d597fd029df"} Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.857432 4754 generic.go:334] "Generic (PLEG): container finished" podID="8f2d4813-8e67-4e79-8aad-5519f6355ced" containerID="129073840aa223733a780a4be753edfc279bc404440474759f083d597fd029df" exitCode=0 Oct 11 03:22:50 crc kubenswrapper[4754]: I1011 03:22:50.859606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"955e7c85-4242-453b-b23d-432057450e28","Type":"ContainerStarted","Data":"0919b403a680a2dac758d230433a87993acbf2d46c9a49df343d1be435ddb550"} Oct 11 03:22:51 crc kubenswrapper[4754]: I1011 03:22:51.098827 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="976f8dd9-0b51-4b45-9192-9ddf098acd48" path="/var/lib/kubelet/pods/976f8dd9-0b51-4b45-9192-9ddf098acd48/volumes" Oct 11 03:22:51 crc kubenswrapper[4754]: I1011 03:22:51.339591 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 11 03:22:51 crc kubenswrapper[4754]: I1011 03:22:51.869690 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"955e7c85-4242-453b-b23d-432057450e28","Type":"ContainerStarted","Data":"533374066f39114c99f70e7d8369cb82bdcabe67742158a0ab543e1db060194a"} Oct 11 03:22:51 crc kubenswrapper[4754]: I1011 03:22:51.870217 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 11 03:22:51 crc kubenswrapper[4754]: I1011 03:22:51.899873 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.534128243 podStartE2EDuration="2.899857486s" podCreationTimestamp="2025-10-11 03:22:49 +0000 UTC" firstStartedPulling="2025-10-11 03:22:50.719992424 +0000 UTC m=+1018.278937209" lastFinishedPulling="2025-10-11 03:22:51.085721667 +0000 UTC m=+1018.644666452" observedRunningTime="2025-10-11 03:22:51.896712058 +0000 UTC m=+1019.455656843" watchObservedRunningTime="2025-10-11 03:22:51.899857486 +0000 UTC m=+1019.458802271" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.352573 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.459439 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rxkh\" (UniqueName: \"kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh\") pod \"8f2d4813-8e67-4e79-8aad-5519f6355ced\" (UID: \"8f2d4813-8e67-4e79-8aad-5519f6355ced\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.464728 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh" (OuterVolumeSpecName: "kube-api-access-7rxkh") pod "8f2d4813-8e67-4e79-8aad-5519f6355ced" (UID: "8f2d4813-8e67-4e79-8aad-5519f6355ced"). InnerVolumeSpecName "kube-api-access-7rxkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.533275 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.536348 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.541945 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.561329 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rxkh\" (UniqueName: \"kubernetes.io/projected/8f2d4813-8e67-4e79-8aad-5519f6355ced-kube-api-access-7rxkh\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.662944 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m5kz\" (UniqueName: \"kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz\") pod \"b3625e97-9933-4bf5-9253-48a0465301fd\" (UID: \"b3625e97-9933-4bf5-9253-48a0465301fd\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663071 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663168 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663204 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n22f4\" (UniqueName: \"kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663243 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663294 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663331 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle\") pod \"a7acd45c-a0be-46e0-9834-f23183f0673a\" (UID: \"a7acd45c-a0be-46e0-9834-f23183f0673a\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.663353 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knshc\" (UniqueName: \"kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc\") pod \"3a5329e2-6d7f-477c-9496-347d8a49f4df\" (UID: \"3a5329e2-6d7f-477c-9496-347d8a49f4df\") " Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.666481 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.668194 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts" (OuterVolumeSpecName: "scripts") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.673312 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.690264 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc" (OuterVolumeSpecName: "kube-api-access-knshc") pod "3a5329e2-6d7f-477c-9496-347d8a49f4df" (UID: "3a5329e2-6d7f-477c-9496-347d8a49f4df"). InnerVolumeSpecName "kube-api-access-knshc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.690405 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4" (OuterVolumeSpecName: "kube-api-access-n22f4") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "kube-api-access-n22f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.690480 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz" (OuterVolumeSpecName: "kube-api-access-2m5kz") pod "b3625e97-9933-4bf5-9253-48a0465301fd" (UID: "b3625e97-9933-4bf5-9253-48a0465301fd"). InnerVolumeSpecName "kube-api-access-2m5kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.746138 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.766929 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.766978 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n22f4\" (UniqueName: \"kubernetes.io/projected/a7acd45c-a0be-46e0-9834-f23183f0673a-kube-api-access-n22f4\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.766989 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7acd45c-a0be-46e0-9834-f23183f0673a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.767007 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.767017 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knshc\" (UniqueName: \"kubernetes.io/projected/3a5329e2-6d7f-477c-9496-347d8a49f4df-kube-api-access-knshc\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.767026 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m5kz\" (UniqueName: \"kubernetes.io/projected/b3625e97-9933-4bf5-9253-48a0465301fd-kube-api-access-2m5kz\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.767034 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.794885 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.832766 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data" (OuterVolumeSpecName: "config-data") pod "a7acd45c-a0be-46e0-9834-f23183f0673a" (UID: "a7acd45c-a0be-46e0-9834-f23183f0673a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.868276 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.868305 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acd45c-a0be-46e0-9834-f23183f0673a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.878524 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkp24" event={"ID":"8f2d4813-8e67-4e79-8aad-5519f6355ced","Type":"ContainerDied","Data":"67b1f40eb624ef880542f556874fe73cf20957a8fa2db23314cbdc5a7cdabcc6"} Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.878559 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67b1f40eb624ef880542f556874fe73cf20957a8fa2db23314cbdc5a7cdabcc6" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.878579 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkp24" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.879921 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rwhqn" event={"ID":"b3625e97-9933-4bf5-9253-48a0465301fd","Type":"ContainerDied","Data":"8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc"} Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.879943 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rwhqn" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.879956 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eaf8687c8a398b5bb08a94eb03f433ea20efe08d90d9a29b8e5281822d617fc" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.882029 4754 generic.go:334] "Generic (PLEG): container finished" podID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerID="8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f" exitCode=0 Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.882070 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerDied","Data":"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f"} Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.882089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7acd45c-a0be-46e0-9834-f23183f0673a","Type":"ContainerDied","Data":"6616035a579ee8fd099fb5fa89db8ebdc3c306ec02cd8318271d40295068ed2f"} Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.882106 4754 scope.go:117] "RemoveContainer" containerID="0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.882212 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.889001 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j2hk9" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.892357 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j2hk9" event={"ID":"3a5329e2-6d7f-477c-9496-347d8a49f4df","Type":"ContainerDied","Data":"47f8d68859afd54bc172c42ed74f8fe0c442710cd17909c6f2436ded02985b23"} Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.892410 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47f8d68859afd54bc172c42ed74f8fe0c442710cd17909c6f2436ded02985b23" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.917614 4754 scope.go:117] "RemoveContainer" containerID="4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.931710 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.936662 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.948243 4754 scope.go:117] "RemoveContainer" containerID="8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.956288 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.956839 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="sg-core" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.956861 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="sg-core" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.956881 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-central-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.956889 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-central-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.956901 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="proxy-httpd" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.956910 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="proxy-httpd" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.956935 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3625e97-9933-4bf5-9253-48a0465301fd" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.956943 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3625e97-9933-4bf5-9253-48a0465301fd" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.956956 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5329e2-6d7f-477c-9496-347d8a49f4df" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963063 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5329e2-6d7f-477c-9496-347d8a49f4df" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.963099 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2d4813-8e67-4e79-8aad-5519f6355ced" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963110 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2d4813-8e67-4e79-8aad-5519f6355ced" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: E1011 03:22:52.963138 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-notification-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963148 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-notification-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963417 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3625e97-9933-4bf5-9253-48a0465301fd" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963440 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2d4813-8e67-4e79-8aad-5519f6355ced" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963452 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-notification-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963461 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5329e2-6d7f-477c-9496-347d8a49f4df" containerName="mariadb-database-create" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963473 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="ceilometer-central-agent" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963488 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="sg-core" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.963501 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" containerName="proxy-httpd" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.965192 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.967365 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.967679 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.981388 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:22:52 crc kubenswrapper[4754]: I1011 03:22:52.985562 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.003249 4754 scope.go:117] "RemoveContainer" containerID="d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.022523 4754 scope.go:117] "RemoveContainer" containerID="0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4" Oct 11 03:22:53 crc kubenswrapper[4754]: E1011 03:22:53.022925 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4\": container with ID starting with 0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4 not found: ID does not exist" containerID="0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.022957 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4"} err="failed to get container status \"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4\": rpc error: code = NotFound desc = could not find container \"0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4\": container with ID starting with 0a84f85ffe50e702dce6c787395c61007d95179c770666d728dce5afd854a0f4 not found: ID does not exist" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.023012 4754 scope.go:117] "RemoveContainer" containerID="4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b" Oct 11 03:22:53 crc kubenswrapper[4754]: E1011 03:22:53.023399 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b\": container with ID starting with 4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b not found: ID does not exist" containerID="4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.023449 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b"} err="failed to get container status \"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b\": rpc error: code = NotFound desc = could not find container \"4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b\": container with ID starting with 4062b391a6c14a32cb598ccbc962c8658905b2d141e02dffdd62e3fe2acfc69b not found: ID does not exist" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.023589 4754 scope.go:117] "RemoveContainer" containerID="8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f" Oct 11 03:22:53 crc kubenswrapper[4754]: E1011 03:22:53.023903 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f\": container with ID starting with 8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f not found: ID does not exist" containerID="8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.023978 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f"} err="failed to get container status \"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f\": rpc error: code = NotFound desc = could not find container \"8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f\": container with ID starting with 8d5992442b8c03e5fe9898da97b411d9e1ff06054c900574891d6493b401a32f not found: ID does not exist" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.024018 4754 scope.go:117] "RemoveContainer" containerID="d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c" Oct 11 03:22:53 crc kubenswrapper[4754]: E1011 03:22:53.024364 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c\": container with ID starting with d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c not found: ID does not exist" containerID="d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.024426 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c"} err="failed to get container status \"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c\": rpc error: code = NotFound desc = could not find container \"d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c\": container with ID starting with d96201f76909661351d4282470858fc051cff15e1ae58c800299c3f5fabb4e9c not found: ID does not exist" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071667 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071708 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071777 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071836 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnqb6\" (UniqueName: \"kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.071885 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.072257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.105756 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7acd45c-a0be-46e0-9834-f23183f0673a" path="/var/lib/kubelet/pods/a7acd45c-a0be-46e0-9834-f23183f0673a/volumes" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173304 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173362 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173400 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173446 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173471 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnqb6\" (UniqueName: \"kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173500 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173707 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.173919 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.175029 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.175817 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.175914 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.177018 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.186802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.187098 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.187341 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.187616 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.191740 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnqb6\" (UniqueName: \"kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6\") pod \"ceilometer-0\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.296661 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:22:53 crc kubenswrapper[4754]: W1011 03:22:53.765277 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3868a10f_655c_4f4e_a765_e39e4e297cb8.slice/crio-9fe399e41a43a3750fa8e5fcfd1f7ad3a4ea4d30dcd0c0303d1bd53b799f1f1c WatchSource:0}: Error finding container 9fe399e41a43a3750fa8e5fcfd1f7ad3a4ea4d30dcd0c0303d1bd53b799f1f1c: Status 404 returned error can't find the container with id 9fe399e41a43a3750fa8e5fcfd1f7ad3a4ea4d30dcd0c0303d1bd53b799f1f1c Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.770032 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:22:53 crc kubenswrapper[4754]: I1011 03:22:53.909271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerStarted","Data":"9fe399e41a43a3750fa8e5fcfd1f7ad3a4ea4d30dcd0c0303d1bd53b799f1f1c"} Oct 11 03:22:55 crc kubenswrapper[4754]: I1011 03:22:55.928077 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerStarted","Data":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.236940 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.938744 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d868-account-create-6gct8"] Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.941509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.944247 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.950430 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerStarted","Data":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.950495 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerStarted","Data":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} Oct 11 03:22:56 crc kubenswrapper[4754]: I1011 03:22:56.951318 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d868-account-create-6gct8"] Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.044183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zcf\" (UniqueName: \"kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf\") pod \"nova-api-d868-account-create-6gct8\" (UID: \"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b\") " pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.145755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zcf\" (UniqueName: \"kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf\") pod \"nova-api-d868-account-create-6gct8\" (UID: \"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b\") " pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.175256 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zcf\" (UniqueName: \"kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf\") pod \"nova-api-d868-account-create-6gct8\" (UID: \"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b\") " pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.260194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.776205 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d868-account-create-6gct8"] Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.971002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d868-account-create-6gct8" event={"ID":"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b","Type":"ContainerStarted","Data":"8f88f642291bfe492294f35c68e1d2403be9a8bde3b27e3c9b7bc4a4476747e4"} Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.971336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d868-account-create-6gct8" event={"ID":"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b","Type":"ContainerStarted","Data":"2b1adffb60434414a727d811a822180003203c53fae177ea053b6eb5652ecb44"} Oct 11 03:22:57 crc kubenswrapper[4754]: I1011 03:22:57.987075 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-d868-account-create-6gct8" podStartSLOduration=1.9870540079999999 podStartE2EDuration="1.987054008s" podCreationTimestamp="2025-10-11 03:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:22:57.982283123 +0000 UTC m=+1025.541227918" watchObservedRunningTime="2025-10-11 03:22:57.987054008 +0000 UTC m=+1025.545998813" Oct 11 03:22:58 crc kubenswrapper[4754]: I1011 03:22:58.982339 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" containerID="8f88f642291bfe492294f35c68e1d2403be9a8bde3b27e3c9b7bc4a4476747e4" exitCode=0 Oct 11 03:22:58 crc kubenswrapper[4754]: I1011 03:22:58.982410 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d868-account-create-6gct8" event={"ID":"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b","Type":"ContainerDied","Data":"8f88f642291bfe492294f35c68e1d2403be9a8bde3b27e3c9b7bc4a4476747e4"} Oct 11 03:22:58 crc kubenswrapper[4754]: I1011 03:22:58.986028 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerStarted","Data":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} Oct 11 03:22:58 crc kubenswrapper[4754]: I1011 03:22:58.986328 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:22:59 crc kubenswrapper[4754]: I1011 03:22:59.025185 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6085290690000003 podStartE2EDuration="7.025166629s" podCreationTimestamp="2025-10-11 03:22:52 +0000 UTC" firstStartedPulling="2025-10-11 03:22:53.76747573 +0000 UTC m=+1021.326420515" lastFinishedPulling="2025-10-11 03:22:58.18411328 +0000 UTC m=+1025.743058075" observedRunningTime="2025-10-11 03:22:59.022608827 +0000 UTC m=+1026.581553632" watchObservedRunningTime="2025-10-11 03:22:59.025166629 +0000 UTC m=+1026.584111414" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.277308 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.342494 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.407280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7zcf\" (UniqueName: \"kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf\") pod \"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b\" (UID: \"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b\") " Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.421200 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf" (OuterVolumeSpecName: "kube-api-access-z7zcf") pod "ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" (UID: "ea9fc5ae-cabb-483c-8aec-d0b5b478d54b"). InnerVolumeSpecName "kube-api-access-z7zcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.510895 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.518517 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7zcf\" (UniqueName: \"kubernetes.io/projected/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b-kube-api-access-z7zcf\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.736543 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.736612 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.736662 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.737376 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:23:00 crc kubenswrapper[4754]: I1011 03:23:00.737432 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494" gracePeriod=600 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.000919 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494" exitCode=0 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.001307 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494"} Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.001337 4754 scope.go:117] "RemoveContainer" containerID="953e3b96a69a99f9a1c7286598ad26645d801a09edf581258071544504f855ac" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.006892 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-central-agent" containerID="cri-o://8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" gracePeriod=30 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.007209 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d868-account-create-6gct8" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.007813 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d868-account-create-6gct8" event={"ID":"ea9fc5ae-cabb-483c-8aec-d0b5b478d54b","Type":"ContainerDied","Data":"2b1adffb60434414a727d811a822180003203c53fae177ea053b6eb5652ecb44"} Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.007833 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="sg-core" containerID="cri-o://4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" gracePeriod=30 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.007875 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b1adffb60434414a727d811a822180003203c53fae177ea053b6eb5652ecb44" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.008306 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="proxy-httpd" containerID="cri-o://a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" gracePeriod=30 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.008349 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-notification-agent" containerID="cri-o://29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" gracePeriod=30 Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.681659 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.838983 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnqb6\" (UniqueName: \"kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.839027 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.839860 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.839892 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.839978 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.839999 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.840065 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.840094 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml\") pod \"3868a10f-655c-4f4e-a765-e39e4e297cb8\" (UID: \"3868a10f-655c-4f4e-a765-e39e4e297cb8\") " Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.841551 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.841638 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.845129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts" (OuterVolumeSpecName: "scripts") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.846597 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6" (OuterVolumeSpecName: "kube-api-access-tnqb6") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "kube-api-access-tnqb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.865283 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.887531 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.909401 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942304 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942349 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942367 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3868a10f-655c-4f4e-a765-e39e4e297cb8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942383 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942403 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942421 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.942438 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnqb6\" (UniqueName: \"kubernetes.io/projected/3868a10f-655c-4f4e-a765-e39e4e297cb8-kube-api-access-tnqb6\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:01 crc kubenswrapper[4754]: I1011 03:23:01.954293 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data" (OuterVolumeSpecName: "config-data") pod "3868a10f-655c-4f4e-a765-e39e4e297cb8" (UID: "3868a10f-655c-4f4e-a765-e39e4e297cb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.015828 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025132 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" exitCode=0 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025164 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" exitCode=2 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025175 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" exitCode=0 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025183 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" exitCode=0 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025212 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerDied","Data":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025238 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerDied","Data":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025248 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerDied","Data":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerDied","Data":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3868a10f-655c-4f4e-a765-e39e4e297cb8","Type":"ContainerDied","Data":"9fe399e41a43a3750fa8e5fcfd1f7ad3a4ea4d30dcd0c0303d1bd53b799f1f1c"} Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025283 4754 scope.go:117] "RemoveContainer" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.025314 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.043594 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3868a10f-655c-4f4e-a765-e39e4e297cb8-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.082822 4754 scope.go:117] "RemoveContainer" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.088941 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.095796 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.107762 4754 scope.go:117] "RemoveContainer" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.108672 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.109142 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-central-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.109229 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-central-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.109292 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-notification-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.109343 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-notification-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.109422 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="proxy-httpd" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.109480 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="proxy-httpd" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.109540 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="sg-core" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.109595 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="sg-core" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.109652 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" containerName="mariadb-account-create" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.109706 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" containerName="mariadb-account-create" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.110008 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-notification-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.110089 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="proxy-httpd" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.110170 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" containerName="mariadb-account-create" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.110249 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="sg-core" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.110319 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" containerName="ceilometer-central-agent" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.115255 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.117769 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.118144 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.118537 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.119105 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.142675 4754 scope.go:117] "RemoveContainer" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.165843 4754 scope.go:117] "RemoveContainer" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.166454 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": container with ID starting with a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706 not found: ID does not exist" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.166506 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} err="failed to get container status \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": rpc error: code = NotFound desc = could not find container \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": container with ID starting with a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.166536 4754 scope.go:117] "RemoveContainer" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.166894 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": container with ID starting with 4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d not found: ID does not exist" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.166932 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} err="failed to get container status \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": rpc error: code = NotFound desc = could not find container \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": container with ID starting with 4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.166978 4754 scope.go:117] "RemoveContainer" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.167229 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": container with ID starting with 29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17 not found: ID does not exist" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167259 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} err="failed to get container status \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": rpc error: code = NotFound desc = could not find container \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": container with ID starting with 29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167273 4754 scope.go:117] "RemoveContainer" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: E1011 03:23:02.167498 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": container with ID starting with 8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48 not found: ID does not exist" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167528 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} err="failed to get container status \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": rpc error: code = NotFound desc = could not find container \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": container with ID starting with 8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167543 4754 scope.go:117] "RemoveContainer" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167743 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} err="failed to get container status \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": rpc error: code = NotFound desc = could not find container \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": container with ID starting with a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.167759 4754 scope.go:117] "RemoveContainer" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168013 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} err="failed to get container status \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": rpc error: code = NotFound desc = could not find container \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": container with ID starting with 4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168032 4754 scope.go:117] "RemoveContainer" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168234 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} err="failed to get container status \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": rpc error: code = NotFound desc = could not find container \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": container with ID starting with 29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168256 4754 scope.go:117] "RemoveContainer" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168536 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} err="failed to get container status \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": rpc error: code = NotFound desc = could not find container \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": container with ID starting with 8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168560 4754 scope.go:117] "RemoveContainer" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168872 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} err="failed to get container status \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": rpc error: code = NotFound desc = could not find container \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": container with ID starting with a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.168892 4754 scope.go:117] "RemoveContainer" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.169192 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} err="failed to get container status \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": rpc error: code = NotFound desc = could not find container \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": container with ID starting with 4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.169297 4754 scope.go:117] "RemoveContainer" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.169671 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} err="failed to get container status \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": rpc error: code = NotFound desc = could not find container \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": container with ID starting with 29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.169696 4754 scope.go:117] "RemoveContainer" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.169923 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} err="failed to get container status \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": rpc error: code = NotFound desc = could not find container \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": container with ID starting with 8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.170060 4754 scope.go:117] "RemoveContainer" containerID="a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.170488 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706"} err="failed to get container status \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": rpc error: code = NotFound desc = could not find container \"a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706\": container with ID starting with a6e79dd4f087246e5cac96e81d300a62040e8e30c7b0bf7934dd8bd25c06a706 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.170515 4754 scope.go:117] "RemoveContainer" containerID="4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.170771 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d"} err="failed to get container status \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": rpc error: code = NotFound desc = could not find container \"4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d\": container with ID starting with 4d67b0146d0143530f8aeb9d821ff00a8914353f3a7b72ad09bae8d969b54f4d not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.170858 4754 scope.go:117] "RemoveContainer" containerID="29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.171215 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17"} err="failed to get container status \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": rpc error: code = NotFound desc = could not find container \"29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17\": container with ID starting with 29370b47118da6065fe814c3a6b9778960ba164150e7def29e0dbb029d455b17 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.171237 4754 scope.go:117] "RemoveContainer" containerID="8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.171493 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48"} err="failed to get container status \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": rpc error: code = NotFound desc = could not find container \"8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48\": container with ID starting with 8b386a5bae6331707e56ad0513563479a0a135c23e62daa4634023147d1c1b48 not found: ID does not exist" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxwd\" (UniqueName: \"kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246689 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246752 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246836 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.246893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348703 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348744 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348793 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxwd\" (UniqueName: \"kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.348871 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.349392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.349439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.356205 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.356256 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-57cbfbbbc9-kgfv2" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.356265 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.357190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.358169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.366000 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxwd\" (UniqueName: \"kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.369676 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data\") pod \"ceilometer-0\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " pod="openstack/ceilometer-0" Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.409596 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.410397 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f8447c5c6-gnvv7" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-api" containerID="cri-o://38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635" gracePeriod=30 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.410836 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f8447c5c6-gnvv7" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-httpd" containerID="cri-o://a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb" gracePeriod=30 Oct 11 03:23:02 crc kubenswrapper[4754]: I1011 03:23:02.434383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:03 crc kubenswrapper[4754]: I1011 03:23:03.000471 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:03 crc kubenswrapper[4754]: W1011 03:23:03.003665 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod041b3ab2_94c1_4c1b_ae2d_78a7d71829b1.slice/crio-d531d323badd73b4ea752a5a049ba2a63281c78fb9b34bbef7eb17b36bb6c4de WatchSource:0}: Error finding container d531d323badd73b4ea752a5a049ba2a63281c78fb9b34bbef7eb17b36bb6c4de: Status 404 returned error can't find the container with id d531d323badd73b4ea752a5a049ba2a63281c78fb9b34bbef7eb17b36bb6c4de Oct 11 03:23:03 crc kubenswrapper[4754]: I1011 03:23:03.033579 4754 generic.go:334] "Generic (PLEG): container finished" podID="0885c413-9d60-465c-95da-965355571d37" containerID="a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb" exitCode=0 Oct 11 03:23:03 crc kubenswrapper[4754]: I1011 03:23:03.033684 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerDied","Data":"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb"} Oct 11 03:23:03 crc kubenswrapper[4754]: I1011 03:23:03.035116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerStarted","Data":"d531d323badd73b4ea752a5a049ba2a63281c78fb9b34bbef7eb17b36bb6c4de"} Oct 11 03:23:03 crc kubenswrapper[4754]: I1011 03:23:03.094929 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3868a10f-655c-4f4e-a765-e39e4e297cb8" path="/var/lib/kubelet/pods/3868a10f-655c-4f4e-a765-e39e4e297cb8/volumes" Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.045815 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerStarted","Data":"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8"} Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.700399 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.778321 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.899505 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config\") pod \"0885c413-9d60-465c-95da-965355571d37\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.899843 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config\") pod \"0885c413-9d60-465c-95da-965355571d37\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.899900 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q8gp\" (UniqueName: \"kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp\") pod \"0885c413-9d60-465c-95da-965355571d37\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.899926 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs\") pod \"0885c413-9d60-465c-95da-965355571d37\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.900000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle\") pod \"0885c413-9d60-465c-95da-965355571d37\" (UID: \"0885c413-9d60-465c-95da-965355571d37\") " Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.904618 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0885c413-9d60-465c-95da-965355571d37" (UID: "0885c413-9d60-465c-95da-965355571d37"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:04 crc kubenswrapper[4754]: I1011 03:23:04.914305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp" (OuterVolumeSpecName: "kube-api-access-9q8gp") pod "0885c413-9d60-465c-95da-965355571d37" (UID: "0885c413-9d60-465c-95da-965355571d37"). InnerVolumeSpecName "kube-api-access-9q8gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.001854 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.001887 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q8gp\" (UniqueName: \"kubernetes.io/projected/0885c413-9d60-465c-95da-965355571d37-kube-api-access-9q8gp\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.002735 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config" (OuterVolumeSpecName: "config") pod "0885c413-9d60-465c-95da-965355571d37" (UID: "0885c413-9d60-465c-95da-965355571d37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.019418 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0885c413-9d60-465c-95da-965355571d37" (UID: "0885c413-9d60-465c-95da-965355571d37"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.022136 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0885c413-9d60-465c-95da-965355571d37" (UID: "0885c413-9d60-465c-95da-965355571d37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.057030 4754 generic.go:334] "Generic (PLEG): container finished" podID="0885c413-9d60-465c-95da-965355571d37" containerID="38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635" exitCode=0 Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.057090 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f8447c5c6-gnvv7" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.057097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerDied","Data":"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635"} Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.057220 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f8447c5c6-gnvv7" event={"ID":"0885c413-9d60-465c-95da-965355571d37","Type":"ContainerDied","Data":"8a3eb65d0183855ccb807e60181d148c47f90ff7867ebdebc8bb313c87a57de9"} Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.057257 4754 scope.go:117] "RemoveContainer" containerID="a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.059868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerStarted","Data":"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1"} Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.059898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerStarted","Data":"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83"} Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.082585 4754 scope.go:117] "RemoveContainer" containerID="38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.103225 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.103266 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.103278 4754 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0885c413-9d60-465c-95da-965355571d37-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.116070 4754 scope.go:117] "RemoveContainer" containerID="a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb" Oct 11 03:23:05 crc kubenswrapper[4754]: E1011 03:23:05.116391 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb\": container with ID starting with a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb not found: ID does not exist" containerID="a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.116429 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb"} err="failed to get container status \"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb\": rpc error: code = NotFound desc = could not find container \"a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb\": container with ID starting with a10bb6840e72b8d037548a6a5cb86421201d6e936d199ecb40d9350a1f7b05eb not found: ID does not exist" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.116451 4754 scope.go:117] "RemoveContainer" containerID="38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635" Oct 11 03:23:05 crc kubenswrapper[4754]: E1011 03:23:05.116649 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635\": container with ID starting with 38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635 not found: ID does not exist" containerID="38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.116670 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635"} err="failed to get container status \"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635\": rpc error: code = NotFound desc = could not find container \"38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635\": container with ID starting with 38bc47b139280e0f33b93c6284b130fe9ff460675428e97b9eefd44c21720635 not found: ID does not exist" Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.118335 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:23:05 crc kubenswrapper[4754]: I1011 03:23:05.121399 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7f8447c5c6-gnvv7"] Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.079990 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerStarted","Data":"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35"} Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.080854 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-central-agent" containerID="cri-o://eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8" gracePeriod=30 Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.081101 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.081430 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="sg-core" containerID="cri-o://e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1" gracePeriod=30 Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.081521 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="proxy-httpd" containerID="cri-o://0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35" gracePeriod=30 Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.081706 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-notification-agent" containerID="cri-o://178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83" gracePeriod=30 Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.098987 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0885c413-9d60-465c-95da-965355571d37" path="/var/lib/kubelet/pods/0885c413-9d60-465c-95da-965355571d37/volumes" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.122486 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.059709089 podStartE2EDuration="5.122463636s" podCreationTimestamp="2025-10-11 03:23:02 +0000 UTC" firstStartedPulling="2025-10-11 03:23:03.006251005 +0000 UTC m=+1030.565195780" lastFinishedPulling="2025-10-11 03:23:06.069005542 +0000 UTC m=+1033.627950327" observedRunningTime="2025-10-11 03:23:07.111680452 +0000 UTC m=+1034.670625237" watchObservedRunningTime="2025-10-11 03:23:07.122463636 +0000 UTC m=+1034.681408431" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.199763 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-bc5d-account-create-qt2kw"] Oct 11 03:23:07 crc kubenswrapper[4754]: E1011 03:23:07.208375 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-api" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.208402 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-api" Oct 11 03:23:07 crc kubenswrapper[4754]: E1011 03:23:07.208412 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-httpd" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.208420 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-httpd" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.210627 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-api" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.210668 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0885c413-9d60-465c-95da-965355571d37" containerName="neutron-httpd" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.212468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.214941 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.245939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcr7f\" (UniqueName: \"kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f\") pod \"nova-cell0-bc5d-account-create-qt2kw\" (UID: \"80dd29fb-8f91-4375-97f8-70c9fff47f7f\") " pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.247420 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bc5d-account-create-qt2kw"] Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.348164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcr7f\" (UniqueName: \"kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f\") pod \"nova-cell0-bc5d-account-create-qt2kw\" (UID: \"80dd29fb-8f91-4375-97f8-70c9fff47f7f\") " pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.351364 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-6de7-account-create-d2fp2"] Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.355827 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.359237 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.366770 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6de7-account-create-d2fp2"] Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.377582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcr7f\" (UniqueName: \"kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f\") pod \"nova-cell0-bc5d-account-create-qt2kw\" (UID: \"80dd29fb-8f91-4375-97f8-70c9fff47f7f\") " pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.449748 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cvxb\" (UniqueName: \"kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb\") pod \"nova-cell1-6de7-account-create-d2fp2\" (UID: \"1c7c15bb-cace-4445-8e0c-d07fe698b3a2\") " pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.551276 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cvxb\" (UniqueName: \"kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb\") pod \"nova-cell1-6de7-account-create-d2fp2\" (UID: \"1c7c15bb-cace-4445-8e0c-d07fe698b3a2\") " pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.554733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.570568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cvxb\" (UniqueName: \"kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb\") pod \"nova-cell1-6de7-account-create-d2fp2\" (UID: \"1c7c15bb-cace-4445-8e0c-d07fe698b3a2\") " pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.669795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:07 crc kubenswrapper[4754]: W1011 03:23:07.992428 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80dd29fb_8f91_4375_97f8_70c9fff47f7f.slice/crio-babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505 WatchSource:0}: Error finding container babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505: Status 404 returned error can't find the container with id babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505 Oct 11 03:23:07 crc kubenswrapper[4754]: I1011 03:23:07.999882 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bc5d-account-create-qt2kw"] Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096702 4754 generic.go:334] "Generic (PLEG): container finished" podID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerID="0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35" exitCode=0 Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096735 4754 generic.go:334] "Generic (PLEG): container finished" podID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerID="e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1" exitCode=2 Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096744 4754 generic.go:334] "Generic (PLEG): container finished" podID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerID="178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83" exitCode=0 Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096787 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerDied","Data":"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35"} Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096814 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerDied","Data":"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1"} Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.096825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerDied","Data":"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83"} Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.098607 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" event={"ID":"80dd29fb-8f91-4375-97f8-70c9fff47f7f","Type":"ContainerStarted","Data":"babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505"} Oct 11 03:23:08 crc kubenswrapper[4754]: I1011 03:23:08.101983 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6de7-account-create-d2fp2"] Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.107922 4754 generic.go:334] "Generic (PLEG): container finished" podID="1c7c15bb-cace-4445-8e0c-d07fe698b3a2" containerID="c3fe91e98bcfa01925eb2916edbbae71a896560ec239041be0266feb1631fe6c" exitCode=0 Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.107986 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6de7-account-create-d2fp2" event={"ID":"1c7c15bb-cace-4445-8e0c-d07fe698b3a2","Type":"ContainerDied","Data":"c3fe91e98bcfa01925eb2916edbbae71a896560ec239041be0266feb1631fe6c"} Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.108042 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6de7-account-create-d2fp2" event={"ID":"1c7c15bb-cace-4445-8e0c-d07fe698b3a2","Type":"ContainerStarted","Data":"8076f3e8620c1f4b0b7ea0f6dc8f4c8c107393287a703dbacb6966a57bbc27d3"} Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.109419 4754 generic.go:334] "Generic (PLEG): container finished" podID="80dd29fb-8f91-4375-97f8-70c9fff47f7f" containerID="4caeb350ec2b7d9b8a3111468651e4cee18a21ea1846dc2d8151b23ad0f9c2b7" exitCode=0 Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.109452 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" event={"ID":"80dd29fb-8f91-4375-97f8-70c9fff47f7f","Type":"ContainerDied","Data":"4caeb350ec2b7d9b8a3111468651e4cee18a21ea1846dc2d8151b23ad0f9c2b7"} Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.576738 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683353 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfxwd\" (UniqueName: \"kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683389 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683452 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683479 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683501 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683529 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.683600 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd\") pod \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\" (UID: \"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1\") " Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.684449 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.685097 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.692219 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd" (OuterVolumeSpecName: "kube-api-access-zfxwd") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "kube-api-access-zfxwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.692535 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts" (OuterVolumeSpecName: "scripts") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.712197 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.740946 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.760141 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785283 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785310 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785321 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785330 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785340 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785349 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfxwd\" (UniqueName: \"kubernetes.io/projected/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-kube-api-access-zfxwd\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.785358 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.786515 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data" (OuterVolumeSpecName: "config-data") pod "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" (UID: "041b3ab2-94c1-4c1b-ae2d-78a7d71829b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:09 crc kubenswrapper[4754]: I1011 03:23:09.887160 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.122059 4754 generic.go:334] "Generic (PLEG): container finished" podID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerID="eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8" exitCode=0 Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.122134 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.122135 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerDied","Data":"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8"} Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.122215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"041b3ab2-94c1-4c1b-ae2d-78a7d71829b1","Type":"ContainerDied","Data":"d531d323badd73b4ea752a5a049ba2a63281c78fb9b34bbef7eb17b36bb6c4de"} Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.122249 4754 scope.go:117] "RemoveContainer" containerID="0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.169676 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.171402 4754 scope.go:117] "RemoveContainer" containerID="e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.238036 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.260744 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.261241 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="sg-core" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261264 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="sg-core" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.261278 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-notification-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261285 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-notification-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.261304 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="proxy-httpd" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261311 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="proxy-httpd" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.261332 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-central-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261339 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-central-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261546 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-central-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261568 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="proxy-httpd" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261576 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="ceilometer-notification-agent" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.261587 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" containerName="sg-core" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.264992 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.273089 4754 scope.go:117] "RemoveContainer" containerID="178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.273482 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.273937 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.274157 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.274303 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.307241 4754 scope.go:117] "RemoveContainer" containerID="eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.328130 4754 scope.go:117] "RemoveContainer" containerID="0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.328595 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35\": container with ID starting with 0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35 not found: ID does not exist" containerID="0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.328651 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35"} err="failed to get container status \"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35\": rpc error: code = NotFound desc = could not find container \"0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35\": container with ID starting with 0e16ce0f14440178899b0b9554a77196f265ba8ac80b98ec30c73c9a2a843e35 not found: ID does not exist" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.328680 4754 scope.go:117] "RemoveContainer" containerID="e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.329330 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1\": container with ID starting with e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1 not found: ID does not exist" containerID="e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.329363 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1"} err="failed to get container status \"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1\": rpc error: code = NotFound desc = could not find container \"e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1\": container with ID starting with e17ddb9e2c88083c6cccb9e9491d86288943c8df3b9b32a423aec717258bc0a1 not found: ID does not exist" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.329386 4754 scope.go:117] "RemoveContainer" containerID="178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.329590 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83\": container with ID starting with 178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83 not found: ID does not exist" containerID="178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.329608 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83"} err="failed to get container status \"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83\": rpc error: code = NotFound desc = could not find container \"178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83\": container with ID starting with 178e27d614c7176d133360b505912f46d1d011b8d6a235bb0dfc7ae94e7f5c83 not found: ID does not exist" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.329619 4754 scope.go:117] "RemoveContainer" containerID="eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8" Oct 11 03:23:10 crc kubenswrapper[4754]: E1011 03:23:10.329794 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8\": container with ID starting with eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8 not found: ID does not exist" containerID="eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.329810 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8"} err="failed to get container status \"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8\": rpc error: code = NotFound desc = could not find container \"eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8\": container with ID starting with eaea5a8770d9c7f4977bc4fb072ce73356d9f9bc7f6d2d982186aee80d8eeee8 not found: ID does not exist" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398573 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398744 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398796 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398819 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.398992 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzwt9\" (UniqueName: \"kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzwt9\" (UniqueName: \"kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500821 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500874 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500896 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500931 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.500946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.509260 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.509422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.518356 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.518605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.519897 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.519983 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.521152 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.524989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzwt9\" (UniqueName: \"kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9\") pod \"ceilometer-0\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.596683 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.602791 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.605628 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.703908 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcr7f\" (UniqueName: \"kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f\") pod \"80dd29fb-8f91-4375-97f8-70c9fff47f7f\" (UID: \"80dd29fb-8f91-4375-97f8-70c9fff47f7f\") " Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.704450 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cvxb\" (UniqueName: \"kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb\") pod \"1c7c15bb-cace-4445-8e0c-d07fe698b3a2\" (UID: \"1c7c15bb-cace-4445-8e0c-d07fe698b3a2\") " Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.708630 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f" (OuterVolumeSpecName: "kube-api-access-kcr7f") pod "80dd29fb-8f91-4375-97f8-70c9fff47f7f" (UID: "80dd29fb-8f91-4375-97f8-70c9fff47f7f"). InnerVolumeSpecName "kube-api-access-kcr7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.709266 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb" (OuterVolumeSpecName: "kube-api-access-6cvxb") pod "1c7c15bb-cace-4445-8e0c-d07fe698b3a2" (UID: "1c7c15bb-cace-4445-8e0c-d07fe698b3a2"). InnerVolumeSpecName "kube-api-access-6cvxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.806522 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcr7f\" (UniqueName: \"kubernetes.io/projected/80dd29fb-8f91-4375-97f8-70c9fff47f7f-kube-api-access-kcr7f\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:10 crc kubenswrapper[4754]: I1011 03:23:10.806552 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cvxb\" (UniqueName: \"kubernetes.io/projected/1c7c15bb-cace-4445-8e0c-d07fe698b3a2-kube-api-access-6cvxb\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:11 crc kubenswrapper[4754]: W1011 03:23:11.090542 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-2ed8321b57c4b4bd505a5db05b5ffdc581e89ffdb6a8090ac638e743cb0ed658 WatchSource:0}: Error finding container 2ed8321b57c4b4bd505a5db05b5ffdc581e89ffdb6a8090ac638e743cb0ed658: Status 404 returned error can't find the container with id 2ed8321b57c4b4bd505a5db05b5ffdc581e89ffdb6a8090ac638e743cb0ed658 Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.095448 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041b3ab2-94c1-4c1b-ae2d-78a7d71829b1" path="/var/lib/kubelet/pods/041b3ab2-94c1-4c1b-ae2d-78a7d71829b1/volumes" Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.104402 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.130296 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerStarted","Data":"2ed8321b57c4b4bd505a5db05b5ffdc581e89ffdb6a8090ac638e743cb0ed658"} Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.134999 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6de7-account-create-d2fp2" Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.135002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6de7-account-create-d2fp2" event={"ID":"1c7c15bb-cace-4445-8e0c-d07fe698b3a2","Type":"ContainerDied","Data":"8076f3e8620c1f4b0b7ea0f6dc8f4c8c107393287a703dbacb6966a57bbc27d3"} Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.135044 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8076f3e8620c1f4b0b7ea0f6dc8f4c8c107393287a703dbacb6966a57bbc27d3" Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.136795 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" event={"ID":"80dd29fb-8f91-4375-97f8-70c9fff47f7f","Type":"ContainerDied","Data":"babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505"} Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.136825 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="babc700fb75592d279ebefe85743a3cd726cc76ef952a75b164e0ad5319b1505" Oct 11 03:23:11 crc kubenswrapper[4754]: I1011 03:23:11.136855 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bc5d-account-create-qt2kw" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.152767 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerStarted","Data":"ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3"} Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.416452 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2grg7"] Oct 11 03:23:12 crc kubenswrapper[4754]: E1011 03:23:12.416772 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80dd29fb-8f91-4375-97f8-70c9fff47f7f" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.416789 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="80dd29fb-8f91-4375-97f8-70c9fff47f7f" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: E1011 03:23:12.416807 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7c15bb-cace-4445-8e0c-d07fe698b3a2" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.416813 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7c15bb-cace-4445-8e0c-d07fe698b3a2" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.417000 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7c15bb-cace-4445-8e0c-d07fe698b3a2" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.417028 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="80dd29fb-8f91-4375-97f8-70c9fff47f7f" containerName="mariadb-account-create" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.417522 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.421355 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.421532 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bvpqg" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.421604 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.433366 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2grg7"] Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.463827 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.463860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.464007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr9k6\" (UniqueName: \"kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.464039 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.565195 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.565436 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.565568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr9k6\" (UniqueName: \"kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.565673 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.568823 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.569748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.580875 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.581064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr9k6\" (UniqueName: \"kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6\") pod \"nova-cell0-conductor-db-sync-2grg7\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:12 crc kubenswrapper[4754]: I1011 03:23:12.730907 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:13 crc kubenswrapper[4754]: I1011 03:23:13.166046 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerStarted","Data":"5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316"} Oct 11 03:23:13 crc kubenswrapper[4754]: I1011 03:23:13.166395 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerStarted","Data":"a288bebe112d781f81fad12fbbe9feca13352223f08d90ad5c5c053966be1c91"} Oct 11 03:23:13 crc kubenswrapper[4754]: I1011 03:23:13.168701 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2grg7"] Oct 11 03:23:14 crc kubenswrapper[4754]: I1011 03:23:14.178538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2grg7" event={"ID":"a5708fd0-491c-4117-a569-8c183034a478","Type":"ContainerStarted","Data":"ec8d92022b33f7e61c8d29a61cc3beb6b5a3dcdbc574b0e14242f082d274b2c1"} Oct 11 03:23:15 crc kubenswrapper[4754]: I1011 03:23:15.189997 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerStarted","Data":"94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d"} Oct 11 03:23:15 crc kubenswrapper[4754]: I1011 03:23:15.190352 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:23:15 crc kubenswrapper[4754]: I1011 03:23:15.210292 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.081777502 podStartE2EDuration="5.210272964s" podCreationTimestamp="2025-10-11 03:23:10 +0000 UTC" firstStartedPulling="2025-10-11 03:23:11.092994595 +0000 UTC m=+1038.651939380" lastFinishedPulling="2025-10-11 03:23:14.221490057 +0000 UTC m=+1041.780434842" observedRunningTime="2025-10-11 03:23:15.208645048 +0000 UTC m=+1042.767589853" watchObservedRunningTime="2025-10-11 03:23:15.210272964 +0000 UTC m=+1042.769217749" Oct 11 03:23:20 crc kubenswrapper[4754]: I1011 03:23:20.230882 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2grg7" event={"ID":"a5708fd0-491c-4117-a569-8c183034a478","Type":"ContainerStarted","Data":"bf61717581adc95c2a00a63f6c90644fce0059e268b08e9d7f9a2489429a64fe"} Oct 11 03:23:20 crc kubenswrapper[4754]: I1011 03:23:20.251569 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2grg7" podStartSLOduration=2.030847905 podStartE2EDuration="8.251546195s" podCreationTimestamp="2025-10-11 03:23:12 +0000 UTC" firstStartedPulling="2025-10-11 03:23:13.173307691 +0000 UTC m=+1040.732252476" lastFinishedPulling="2025-10-11 03:23:19.394005981 +0000 UTC m=+1046.952950766" observedRunningTime="2025-10-11 03:23:20.246098051 +0000 UTC m=+1047.805042856" watchObservedRunningTime="2025-10-11 03:23:20.251546195 +0000 UTC m=+1047.810490980" Oct 11 03:23:30 crc kubenswrapper[4754]: I1011 03:23:30.366855 4754 generic.go:334] "Generic (PLEG): container finished" podID="a5708fd0-491c-4117-a569-8c183034a478" containerID="bf61717581adc95c2a00a63f6c90644fce0059e268b08e9d7f9a2489429a64fe" exitCode=0 Oct 11 03:23:30 crc kubenswrapper[4754]: I1011 03:23:30.367014 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2grg7" event={"ID":"a5708fd0-491c-4117-a569-8c183034a478","Type":"ContainerDied","Data":"bf61717581adc95c2a00a63f6c90644fce0059e268b08e9d7f9a2489429a64fe"} Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.713479 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.809883 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle\") pod \"a5708fd0-491c-4117-a569-8c183034a478\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.810056 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data\") pod \"a5708fd0-491c-4117-a569-8c183034a478\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.810153 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts\") pod \"a5708fd0-491c-4117-a569-8c183034a478\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.810242 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr9k6\" (UniqueName: \"kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6\") pod \"a5708fd0-491c-4117-a569-8c183034a478\" (UID: \"a5708fd0-491c-4117-a569-8c183034a478\") " Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.815696 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts" (OuterVolumeSpecName: "scripts") pod "a5708fd0-491c-4117-a569-8c183034a478" (UID: "a5708fd0-491c-4117-a569-8c183034a478"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.816070 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6" (OuterVolumeSpecName: "kube-api-access-gr9k6") pod "a5708fd0-491c-4117-a569-8c183034a478" (UID: "a5708fd0-491c-4117-a569-8c183034a478"). InnerVolumeSpecName "kube-api-access-gr9k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.842647 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data" (OuterVolumeSpecName: "config-data") pod "a5708fd0-491c-4117-a569-8c183034a478" (UID: "a5708fd0-491c-4117-a569-8c183034a478"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.855461 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5708fd0-491c-4117-a569-8c183034a478" (UID: "a5708fd0-491c-4117-a569-8c183034a478"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.912670 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.912704 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr9k6\" (UniqueName: \"kubernetes.io/projected/a5708fd0-491c-4117-a569-8c183034a478-kube-api-access-gr9k6\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.912715 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:31 crc kubenswrapper[4754]: I1011 03:23:31.912726 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5708fd0-491c-4117-a569-8c183034a478-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.387245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2grg7" event={"ID":"a5708fd0-491c-4117-a569-8c183034a478","Type":"ContainerDied","Data":"ec8d92022b33f7e61c8d29a61cc3beb6b5a3dcdbc574b0e14242f082d274b2c1"} Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.387550 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec8d92022b33f7e61c8d29a61cc3beb6b5a3dcdbc574b0e14242f082d274b2c1" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.387459 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2grg7" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.477613 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 03:23:32 crc kubenswrapper[4754]: E1011 03:23:32.477994 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5708fd0-491c-4117-a569-8c183034a478" containerName="nova-cell0-conductor-db-sync" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.478012 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5708fd0-491c-4117-a569-8c183034a478" containerName="nova-cell0-conductor-db-sync" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.478188 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5708fd0-491c-4117-a569-8c183034a478" containerName="nova-cell0-conductor-db-sync" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.478714 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.480524 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.480548 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bvpqg" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.527803 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.527870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.527918 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgf85\" (UniqueName: \"kubernetes.io/projected/198c0494-b8ae-45a5-a666-b60c0de08d0f-kube-api-access-bgf85\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.543006 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.629930 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.630011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.630141 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgf85\" (UniqueName: \"kubernetes.io/projected/198c0494-b8ae-45a5-a666-b60c0de08d0f-kube-api-access-bgf85\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.634428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.636670 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198c0494-b8ae-45a5-a666-b60c0de08d0f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.644846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgf85\" (UniqueName: \"kubernetes.io/projected/198c0494-b8ae-45a5-a666-b60c0de08d0f-kube-api-access-bgf85\") pod \"nova-cell0-conductor-0\" (UID: \"198c0494-b8ae-45a5-a666-b60c0de08d0f\") " pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:32 crc kubenswrapper[4754]: I1011 03:23:32.826399 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:33 crc kubenswrapper[4754]: I1011 03:23:33.258168 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 03:23:33 crc kubenswrapper[4754]: W1011 03:23:33.264200 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod198c0494_b8ae_45a5_a666_b60c0de08d0f.slice/crio-073d8360664d0485298488aaf1754649758fca175d878e4bb8d35dfcd5b003a0 WatchSource:0}: Error finding container 073d8360664d0485298488aaf1754649758fca175d878e4bb8d35dfcd5b003a0: Status 404 returned error can't find the container with id 073d8360664d0485298488aaf1754649758fca175d878e4bb8d35dfcd5b003a0 Oct 11 03:23:33 crc kubenswrapper[4754]: I1011 03:23:33.396511 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"198c0494-b8ae-45a5-a666-b60c0de08d0f","Type":"ContainerStarted","Data":"073d8360664d0485298488aaf1754649758fca175d878e4bb8d35dfcd5b003a0"} Oct 11 03:23:34 crc kubenswrapper[4754]: I1011 03:23:34.407788 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"198c0494-b8ae-45a5-a666-b60c0de08d0f","Type":"ContainerStarted","Data":"3ab848372bc0cce8e75aab097e313a3f2c23ff939179229963af2db621e59352"} Oct 11 03:23:34 crc kubenswrapper[4754]: I1011 03:23:34.408990 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:34 crc kubenswrapper[4754]: I1011 03:23:34.432177 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.432158141 podStartE2EDuration="2.432158141s" podCreationTimestamp="2025-10-11 03:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:34.427832139 +0000 UTC m=+1061.986776964" watchObservedRunningTime="2025-10-11 03:23:34.432158141 +0000 UTC m=+1061.991102926" Oct 11 03:23:40 crc kubenswrapper[4754]: I1011 03:23:40.604304 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 03:23:42 crc kubenswrapper[4754]: I1011 03:23:42.849769 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.278875 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bsdg6"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.280186 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.282203 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.289034 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.302584 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bsdg6"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.432369 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.432477 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.432526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlfv8\" (UniqueName: \"kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.432548 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.471416 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.472935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.476841 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.485380 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.535598 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.535683 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.535719 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlfv8\" (UniqueName: \"kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.535737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.545085 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.545689 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.551989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.571270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlfv8\" (UniqueName: \"kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8\") pod \"nova-cell0-cell-mapping-bsdg6\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.581484 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.583007 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.598263 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.639019 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlkbh\" (UniqueName: \"kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.639062 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.639089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.639133 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.645286 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.646995 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.647323 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.650403 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.690251 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.691314 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.717878 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.742152 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743117 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743173 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw2q2\" (UniqueName: \"kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743199 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743223 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743249 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743287 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdfpz\" (UniqueName: \"kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743324 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743351 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743394 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743414 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlkbh\" (UniqueName: \"kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.743456 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.744417 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.747892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.749187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.767864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlkbh\" (UniqueName: \"kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh\") pod \"nova-metadata-0\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.768714 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.768870 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.778406 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.784143 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.798891 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.805173 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848567 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848840 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdfpz\" (UniqueName: \"kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848879 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848907 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848950 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848980 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.848999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7b8s\" (UniqueName: \"kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849105 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849128 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw2q2\" (UniqueName: \"kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849150 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849652 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849839 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.849857 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.850266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.850333 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.861841 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.862419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.866558 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw2q2\" (UniqueName: \"kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2\") pod \"dnsmasq-dns-566b5b7845-l6xgx\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.866692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdfpz\" (UniqueName: \"kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz\") pod \"nova-api-0\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " pod="openstack/nova-api-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.952352 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954179 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954343 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7b8s\" (UniqueName: \"kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954443 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2tzf\" (UniqueName: \"kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.954470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.959412 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.959480 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.975521 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7b8s\" (UniqueName: \"kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s\") pod \"nova-scheduler-0\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:43 crc kubenswrapper[4754]: I1011 03:23:43.989395 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.055988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.056146 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2tzf\" (UniqueName: \"kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.056171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.061717 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.064549 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.075620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2tzf\" (UniqueName: \"kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf\") pod \"nova-cell1-novncproxy-0\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.163254 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.171197 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.229918 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bsdg6"] Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.311009 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zn2qx"] Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.312496 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.318136 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.318295 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.328618 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zn2qx"] Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.366824 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:44 crc kubenswrapper[4754]: W1011 03:23:44.379302 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode32a00af_7e9c_4e09_a084_ee3cd20cce65.slice/crio-b8c794586c09ebdd367b760e4075c1ccc955e0258c396610d2df3d594171ba95 WatchSource:0}: Error finding container b8c794586c09ebdd367b760e4075c1ccc955e0258c396610d2df3d594171ba95: Status 404 returned error can't find the container with id b8c794586c09ebdd367b760e4075c1ccc955e0258c396610d2df3d594171ba95 Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.466053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.466163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.466217 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.466290 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wks89\" (UniqueName: \"kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.506396 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.518625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bsdg6" event={"ID":"185f8b44-dd99-48d0-940a-7f14690d6728","Type":"ContainerStarted","Data":"75ab5f194447df8a5e6b69fe1edd12c7d18b67205824e54ed79b39ac2fcf0d4d"} Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.519856 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerStarted","Data":"b8c794586c09ebdd367b760e4075c1ccc955e0258c396610d2df3d594171ba95"} Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.572078 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.572190 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.572225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.572278 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wks89\" (UniqueName: \"kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.579873 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.580385 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.580557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.595546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wks89\" (UniqueName: \"kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89\") pod \"nova-cell1-conductor-db-sync-zn2qx\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.618032 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.651241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.723453 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:23:44 crc kubenswrapper[4754]: W1011 03:23:44.732649 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57454bdf_3bd3_444a_938d_f3467c9ec40d.slice/crio-0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803 WatchSource:0}: Error finding container 0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803: Status 404 returned error can't find the container with id 0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803 Oct 11 03:23:44 crc kubenswrapper[4754]: I1011 03:23:44.789252 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:44 crc kubenswrapper[4754]: W1011 03:23:44.795585 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e543740_8abc_41f7_8378_a4d7aaf2ad32.slice/crio-66983ef1e858d02a64f64af7372df0a92860a46e02a7806e9a67d24d95836026 WatchSource:0}: Error finding container 66983ef1e858d02a64f64af7372df0a92860a46e02a7806e9a67d24d95836026: Status 404 returned error can't find the container with id 66983ef1e858d02a64f64af7372df0a92860a46e02a7806e9a67d24d95836026 Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.137820 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zn2qx"] Oct 11 03:23:45 crc kubenswrapper[4754]: W1011 03:23:45.145286 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1639fb4_c6c2_48b4_a13c_6f5166229dac.slice/crio-2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929 WatchSource:0}: Error finding container 2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929: Status 404 returned error can't find the container with id 2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929 Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.533050 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" event={"ID":"a1639fb4-c6c2-48b4-a13c-6f5166229dac","Type":"ContainerStarted","Data":"85a54549196e899d44baab37dac0be4bb16c869a7593d0408d0aaacdea31dcb8"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.533396 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" event={"ID":"a1639fb4-c6c2-48b4-a13c-6f5166229dac","Type":"ContainerStarted","Data":"2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.542493 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57454bdf-3bd3-444a-938d-f3467c9ec40d","Type":"ContainerStarted","Data":"0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.546395 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerStarted","Data":"bb5416b4b1bf54cf40538a7978ba2f6284cd0ff7ed8619e32a9739c4c3b79d5c"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.549672 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e543740-8abc-41f7-8378-a4d7aaf2ad32","Type":"ContainerStarted","Data":"66983ef1e858d02a64f64af7372df0a92860a46e02a7806e9a67d24d95836026"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.551571 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" podStartSLOduration=1.551562057 podStartE2EDuration="1.551562057s" podCreationTimestamp="2025-10-11 03:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:45.550787375 +0000 UTC m=+1073.109732160" watchObservedRunningTime="2025-10-11 03:23:45.551562057 +0000 UTC m=+1073.110506842" Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.559607 4754 generic.go:334] "Generic (PLEG): container finished" podID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerID="d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d" exitCode=0 Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.559691 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" event={"ID":"ab70a33e-1420-4d1d-9af7-b2f93e767536","Type":"ContainerDied","Data":"d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.559712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" event={"ID":"ab70a33e-1420-4d1d-9af7-b2f93e767536","Type":"ContainerStarted","Data":"c5b28be5461fd518d215e9ffb4037b64a24b53d87a5b1c2fd0a21075e77a1488"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.563718 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bsdg6" event={"ID":"185f8b44-dd99-48d0-940a-7f14690d6728","Type":"ContainerStarted","Data":"173034bae9024d38dc086dac8fe19051ee21d16c172abd6e7f17b1f0d75f1e4d"} Oct 11 03:23:45 crc kubenswrapper[4754]: I1011 03:23:45.601688 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bsdg6" podStartSLOduration=2.601666661 podStartE2EDuration="2.601666661s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:45.597824613 +0000 UTC m=+1073.156769408" watchObservedRunningTime="2025-10-11 03:23:45.601666661 +0000 UTC m=+1073.160611446" Oct 11 03:23:46 crc kubenswrapper[4754]: I1011 03:23:46.587949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerStarted","Data":"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f"} Oct 11 03:23:46 crc kubenswrapper[4754]: I1011 03:23:46.590601 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerStarted","Data":"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852"} Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.210892 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.222685 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.601772 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" event={"ID":"ab70a33e-1420-4d1d-9af7-b2f93e767536","Type":"ContainerStarted","Data":"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed"} Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.601923 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.610681 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerStarted","Data":"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f"} Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.616498 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerStarted","Data":"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b"} Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.625582 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" podStartSLOduration=4.625564975 podStartE2EDuration="4.625564975s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:47.618034202 +0000 UTC m=+1075.176978997" watchObservedRunningTime="2025-10-11 03:23:47.625564975 +0000 UTC m=+1075.184509750" Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.640071 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.678510998 podStartE2EDuration="4.640057824s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="2025-10-11 03:23:44.383657232 +0000 UTC m=+1071.942602007" lastFinishedPulling="2025-10-11 03:23:46.345204048 +0000 UTC m=+1073.904148833" observedRunningTime="2025-10-11 03:23:47.6377916 +0000 UTC m=+1075.196736385" watchObservedRunningTime="2025-10-11 03:23:47.640057824 +0000 UTC m=+1075.199002609" Oct 11 03:23:47 crc kubenswrapper[4754]: I1011 03:23:47.663252 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.7732380020000003 podStartE2EDuration="4.663233088s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="2025-10-11 03:23:44.514564147 +0000 UTC m=+1072.073508932" lastFinishedPulling="2025-10-11 03:23:46.404559233 +0000 UTC m=+1073.963504018" observedRunningTime="2025-10-11 03:23:47.655548211 +0000 UTC m=+1075.214492996" watchObservedRunningTime="2025-10-11 03:23:47.663233088 +0000 UTC m=+1075.222177873" Oct 11 03:23:48 crc kubenswrapper[4754]: I1011 03:23:48.625989 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-log" containerID="cri-o://b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" gracePeriod=30 Oct 11 03:23:48 crc kubenswrapper[4754]: I1011 03:23:48.626102 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-metadata" containerID="cri-o://78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" gracePeriod=30 Oct 11 03:23:48 crc kubenswrapper[4754]: I1011 03:23:48.806335 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:23:48 crc kubenswrapper[4754]: I1011 03:23:48.806679 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.100492 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.286742 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data\") pod \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.286826 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlkbh\" (UniqueName: \"kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh\") pod \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.287005 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs\") pod \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.287655 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs" (OuterVolumeSpecName: "logs") pod "e32a00af-7e9c-4e09-a084-ee3cd20cce65" (UID: "e32a00af-7e9c-4e09-a084-ee3cd20cce65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.287774 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle\") pod \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\" (UID: \"e32a00af-7e9c-4e09-a084-ee3cd20cce65\") " Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.288454 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e32a00af-7e9c-4e09-a084-ee3cd20cce65-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.291874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh" (OuterVolumeSpecName: "kube-api-access-zlkbh") pod "e32a00af-7e9c-4e09-a084-ee3cd20cce65" (UID: "e32a00af-7e9c-4e09-a084-ee3cd20cce65"). InnerVolumeSpecName "kube-api-access-zlkbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.312373 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data" (OuterVolumeSpecName: "config-data") pod "e32a00af-7e9c-4e09-a084-ee3cd20cce65" (UID: "e32a00af-7e9c-4e09-a084-ee3cd20cce65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.312695 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e32a00af-7e9c-4e09-a084-ee3cd20cce65" (UID: "e32a00af-7e9c-4e09-a084-ee3cd20cce65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.389996 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.390039 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e32a00af-7e9c-4e09-a084-ee3cd20cce65-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.390058 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlkbh\" (UniqueName: \"kubernetes.io/projected/e32a00af-7e9c-4e09-a084-ee3cd20cce65-kube-api-access-zlkbh\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.652141 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e543740-8abc-41f7-8378-a4d7aaf2ad32","Type":"ContainerStarted","Data":"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5"} Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.657086 4754 generic.go:334] "Generic (PLEG): container finished" podID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerID="78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" exitCode=0 Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.657204 4754 generic.go:334] "Generic (PLEG): container finished" podID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerID="b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" exitCode=143 Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.657152 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.660901 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerDied","Data":"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f"} Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.661022 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerDied","Data":"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852"} Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.661036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e32a00af-7e9c-4e09-a084-ee3cd20cce65","Type":"ContainerDied","Data":"b8c794586c09ebdd367b760e4075c1ccc955e0258c396610d2df3d594171ba95"} Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.661073 4754 scope.go:117] "RemoveContainer" containerID="78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.664801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57454bdf-3bd3-444a-938d-f3467c9ec40d","Type":"ContainerStarted","Data":"68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4"} Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.665016 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="57454bdf-3bd3-444a-938d-f3467c9ec40d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4" gracePeriod=30 Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.691902 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.085791565 podStartE2EDuration="6.691883148s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="2025-10-11 03:23:44.798185942 +0000 UTC m=+1072.357130727" lastFinishedPulling="2025-10-11 03:23:48.404277525 +0000 UTC m=+1075.963222310" observedRunningTime="2025-10-11 03:23:49.677065439 +0000 UTC m=+1077.236010264" watchObservedRunningTime="2025-10-11 03:23:49.691883148 +0000 UTC m=+1077.250827923" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.693565 4754 scope.go:117] "RemoveContainer" containerID="b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.701994 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.034014344 podStartE2EDuration="6.701977233s" podCreationTimestamp="2025-10-11 03:23:43 +0000 UTC" firstStartedPulling="2025-10-11 03:23:44.741818981 +0000 UTC m=+1072.300763766" lastFinishedPulling="2025-10-11 03:23:48.40978187 +0000 UTC m=+1075.968726655" observedRunningTime="2025-10-11 03:23:49.693594136 +0000 UTC m=+1077.252538931" watchObservedRunningTime="2025-10-11 03:23:49.701977233 +0000 UTC m=+1077.260922028" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.719444 4754 scope.go:117] "RemoveContainer" containerID="78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.720189 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:49 crc kubenswrapper[4754]: E1011 03:23:49.720487 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f\": container with ID starting with 78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f not found: ID does not exist" containerID="78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.720522 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f"} err="failed to get container status \"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f\": rpc error: code = NotFound desc = could not find container \"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f\": container with ID starting with 78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f not found: ID does not exist" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.720551 4754 scope.go:117] "RemoveContainer" containerID="b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" Oct 11 03:23:49 crc kubenswrapper[4754]: E1011 03:23:49.721598 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852\": container with ID starting with b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852 not found: ID does not exist" containerID="b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.721622 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852"} err="failed to get container status \"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852\": rpc error: code = NotFound desc = could not find container \"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852\": container with ID starting with b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852 not found: ID does not exist" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.721637 4754 scope.go:117] "RemoveContainer" containerID="78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.721847 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f"} err="failed to get container status \"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f\": rpc error: code = NotFound desc = could not find container \"78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f\": container with ID starting with 78d571e60bb0f7adb28a29ece8b630de2fccd4ba4142116b7e1635e31a3c845f not found: ID does not exist" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.721864 4754 scope.go:117] "RemoveContainer" containerID="b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.722278 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852"} err="failed to get container status \"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852\": rpc error: code = NotFound desc = could not find container \"b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852\": container with ID starting with b0c3adbbb6da7b9b11d45d24dbfd8cb9068952d8739effac4fdcae30031b5852 not found: ID does not exist" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.738145 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.744246 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:49 crc kubenswrapper[4754]: E1011 03:23:49.744827 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-metadata" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.744847 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-metadata" Oct 11 03:23:49 crc kubenswrapper[4754]: E1011 03:23:49.744876 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-log" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.744887 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-log" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.745239 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-metadata" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.745258 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" containerName="nova-metadata-log" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.746716 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.750392 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.750459 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.751703 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.897707 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.897819 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw7rf\" (UniqueName: \"kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.897846 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.897888 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.897919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.999662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw7rf\" (UniqueName: \"kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.999739 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:49 crc kubenswrapper[4754]: I1011 03:23:49.999784 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:49.999836 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:49.999932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.000329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.004829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.006809 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.005461 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.020689 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw7rf\" (UniqueName: \"kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf\") pod \"nova-metadata-0\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.073037 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.606894 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:50 crc kubenswrapper[4754]: I1011 03:23:50.674141 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerStarted","Data":"4fd57cc3c1bdf41203bc29c3bf73e0982c2438170d697e5fa96f56d5deceff3f"} Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.096151 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e32a00af-7e9c-4e09-a084-ee3cd20cce65" path="/var/lib/kubelet/pods/e32a00af-7e9c-4e09-a084-ee3cd20cce65/volumes" Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.698244 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerStarted","Data":"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414"} Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.698554 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerStarted","Data":"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245"} Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.699985 4754 generic.go:334] "Generic (PLEG): container finished" podID="185f8b44-dd99-48d0-940a-7f14690d6728" containerID="173034bae9024d38dc086dac8fe19051ee21d16c172abd6e7f17b1f0d75f1e4d" exitCode=0 Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.700025 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bsdg6" event={"ID":"185f8b44-dd99-48d0-940a-7f14690d6728","Type":"ContainerDied","Data":"173034bae9024d38dc086dac8fe19051ee21d16c172abd6e7f17b1f0d75f1e4d"} Oct 11 03:23:51 crc kubenswrapper[4754]: I1011 03:23:51.724347 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.724326023 podStartE2EDuration="2.724326023s" podCreationTimestamp="2025-10-11 03:23:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:51.719509527 +0000 UTC m=+1079.278454312" watchObservedRunningTime="2025-10-11 03:23:51.724326023 +0000 UTC m=+1079.283270808" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.070542 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.261767 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle\") pod \"185f8b44-dd99-48d0-940a-7f14690d6728\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.261898 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data\") pod \"185f8b44-dd99-48d0-940a-7f14690d6728\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.262049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts\") pod \"185f8b44-dd99-48d0-940a-7f14690d6728\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.262102 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlfv8\" (UniqueName: \"kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8\") pod \"185f8b44-dd99-48d0-940a-7f14690d6728\" (UID: \"185f8b44-dd99-48d0-940a-7f14690d6728\") " Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.267940 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts" (OuterVolumeSpecName: "scripts") pod "185f8b44-dd99-48d0-940a-7f14690d6728" (UID: "185f8b44-dd99-48d0-940a-7f14690d6728"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.268636 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8" (OuterVolumeSpecName: "kube-api-access-rlfv8") pod "185f8b44-dd99-48d0-940a-7f14690d6728" (UID: "185f8b44-dd99-48d0-940a-7f14690d6728"). InnerVolumeSpecName "kube-api-access-rlfv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.293193 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "185f8b44-dd99-48d0-940a-7f14690d6728" (UID: "185f8b44-dd99-48d0-940a-7f14690d6728"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.305006 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data" (OuterVolumeSpecName: "config-data") pod "185f8b44-dd99-48d0-940a-7f14690d6728" (UID: "185f8b44-dd99-48d0-940a-7f14690d6728"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.364132 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.364647 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.364676 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlfv8\" (UniqueName: \"kubernetes.io/projected/185f8b44-dd99-48d0-940a-7f14690d6728-kube-api-access-rlfv8\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.364706 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f8b44-dd99-48d0-940a-7f14690d6728-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.725446 4754 generic.go:334] "Generic (PLEG): container finished" podID="a1639fb4-c6c2-48b4-a13c-6f5166229dac" containerID="85a54549196e899d44baab37dac0be4bb16c869a7593d0408d0aaacdea31dcb8" exitCode=0 Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.725513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" event={"ID":"a1639fb4-c6c2-48b4-a13c-6f5166229dac","Type":"ContainerDied","Data":"85a54549196e899d44baab37dac0be4bb16c869a7593d0408d0aaacdea31dcb8"} Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.727726 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bsdg6" event={"ID":"185f8b44-dd99-48d0-940a-7f14690d6728","Type":"ContainerDied","Data":"75ab5f194447df8a5e6b69fe1edd12c7d18b67205824e54ed79b39ac2fcf0d4d"} Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.727753 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75ab5f194447df8a5e6b69fe1edd12c7d18b67205824e54ed79b39ac2fcf0d4d" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.727809 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bsdg6" Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.912021 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.912295 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-log" containerID="cri-o://07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" gracePeriod=30 Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.912826 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-api" containerID="cri-o://b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" gracePeriod=30 Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.937024 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.937336 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" containerName="nova-scheduler-scheduler" containerID="cri-o://a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5" gracePeriod=30 Oct 11 03:23:53 crc kubenswrapper[4754]: I1011 03:23:53.954119 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.028341 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.028573 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="dnsmasq-dns" containerID="cri-o://0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524" gracePeriod=10 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.048552 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.049762 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-log" containerID="cri-o://b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" gracePeriod=30 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.050219 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-metadata" containerID="cri-o://dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" gracePeriod=30 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.164334 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.171713 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.477748 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.577982 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.587639 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc\") pod \"35a5aafb-45cd-47a0-aae1-e564d0d34848\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.587843 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config\") pod \"35a5aafb-45cd-47a0-aae1-e564d0d34848\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.587896 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpjld\" (UniqueName: \"kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld\") pod \"35a5aafb-45cd-47a0-aae1-e564d0d34848\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.588011 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb\") pod \"35a5aafb-45cd-47a0-aae1-e564d0d34848\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.588039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb\") pod \"35a5aafb-45cd-47a0-aae1-e564d0d34848\" (UID: \"35a5aafb-45cd-47a0-aae1-e564d0d34848\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.593866 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld" (OuterVolumeSpecName: "kube-api-access-wpjld") pod "35a5aafb-45cd-47a0-aae1-e564d0d34848" (UID: "35a5aafb-45cd-47a0-aae1-e564d0d34848"). InnerVolumeSpecName "kube-api-access-wpjld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.597253 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpjld\" (UniqueName: \"kubernetes.io/projected/35a5aafb-45cd-47a0-aae1-e564d0d34848-kube-api-access-wpjld\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.653213 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35a5aafb-45cd-47a0-aae1-e564d0d34848" (UID: "35a5aafb-45cd-47a0-aae1-e564d0d34848"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.659467 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config" (OuterVolumeSpecName: "config") pod "35a5aafb-45cd-47a0-aae1-e564d0d34848" (UID: "35a5aafb-45cd-47a0-aae1-e564d0d34848"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.660518 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.663298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35a5aafb-45cd-47a0-aae1-e564d0d34848" (UID: "35a5aafb-45cd-47a0-aae1-e564d0d34848"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.683768 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35a5aafb-45cd-47a0-aae1-e564d0d34848" (UID: "35a5aafb-45cd-47a0-aae1-e564d0d34848"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698229 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs\") pod \"e0f222a5-c392-4e76-a018-856c6c7bd45d\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698304 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdfpz\" (UniqueName: \"kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz\") pod \"e0f222a5-c392-4e76-a018-856c6c7bd45d\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698375 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data\") pod \"e0f222a5-c392-4e76-a018-856c6c7bd45d\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698480 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle\") pod \"e0f222a5-c392-4e76-a018-856c6c7bd45d\" (UID: \"e0f222a5-c392-4e76-a018-856c6c7bd45d\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698890 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698926 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698953 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.698995 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35a5aafb-45cd-47a0-aae1-e564d0d34848-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.700944 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs" (OuterVolumeSpecName: "logs") pod "e0f222a5-c392-4e76-a018-856c6c7bd45d" (UID: "e0f222a5-c392-4e76-a018-856c6c7bd45d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.701519 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz" (OuterVolumeSpecName: "kube-api-access-kdfpz") pod "e0f222a5-c392-4e76-a018-856c6c7bd45d" (UID: "e0f222a5-c392-4e76-a018-856c6c7bd45d"). InnerVolumeSpecName "kube-api-access-kdfpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.725463 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0f222a5-c392-4e76-a018-856c6c7bd45d" (UID: "e0f222a5-c392-4e76-a018-856c6c7bd45d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.729556 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data" (OuterVolumeSpecName: "config-data") pod "e0f222a5-c392-4e76-a018-856c6c7bd45d" (UID: "e0f222a5-c392-4e76-a018-856c6c7bd45d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.748769 4754 generic.go:334] "Generic (PLEG): container finished" podID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerID="0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524" exitCode=0 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.748812 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" event={"ID":"35a5aafb-45cd-47a0-aae1-e564d0d34848","Type":"ContainerDied","Data":"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.749119 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" event={"ID":"35a5aafb-45cd-47a0-aae1-e564d0d34848","Type":"ContainerDied","Data":"a6a32d3cd441a80d701c57c17a2da1cdf99936e412f077af5c5e1ddb7fbf809f"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.749182 4754 scope.go:117] "RemoveContainer" containerID="0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.748891 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-6dpxh" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752348 4754 generic.go:334] "Generic (PLEG): container finished" podID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerID="dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" exitCode=0 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752370 4754 generic.go:334] "Generic (PLEG): container finished" podID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerID="b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" exitCode=143 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752412 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerDied","Data":"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752431 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerDied","Data":"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752444 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aab1dc13-46c4-4776-a5d0-59092fd5ec78","Type":"ContainerDied","Data":"4fd57cc3c1bdf41203bc29c3bf73e0982c2438170d697e5fa96f56d5deceff3f"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.752412 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754154 4754 generic.go:334] "Generic (PLEG): container finished" podID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerID="b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" exitCode=0 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754179 4754 generic.go:334] "Generic (PLEG): container finished" podID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerID="07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" exitCode=143 Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerDied","Data":"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754344 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerDied","Data":"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0f222a5-c392-4e76-a018-856c6c7bd45d","Type":"ContainerDied","Data":"bb5416b4b1bf54cf40538a7978ba2f6284cd0ff7ed8619e32a9739c4c3b79d5c"} Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.754417 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.786283 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.789802 4754 scope.go:117] "RemoveContainer" containerID="272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.792820 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-6dpxh"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.799931 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw7rf\" (UniqueName: \"kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf\") pod \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.799996 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle\") pod \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800217 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs\") pod \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs\") pod \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800287 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data\") pod \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\" (UID: \"aab1dc13-46c4-4776-a5d0-59092fd5ec78\") " Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800567 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800586 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f222a5-c392-4e76-a018-856c6c7bd45d-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800597 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdfpz\" (UniqueName: \"kubernetes.io/projected/e0f222a5-c392-4e76-a018-856c6c7bd45d-kube-api-access-kdfpz\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.800610 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f222a5-c392-4e76-a018-856c6c7bd45d-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.801206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs" (OuterVolumeSpecName: "logs") pod "aab1dc13-46c4-4776-a5d0-59092fd5ec78" (UID: "aab1dc13-46c4-4776-a5d0-59092fd5ec78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.804103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf" (OuterVolumeSpecName: "kube-api-access-rw7rf") pod "aab1dc13-46c4-4776-a5d0-59092fd5ec78" (UID: "aab1dc13-46c4-4776-a5d0-59092fd5ec78"). InnerVolumeSpecName "kube-api-access-rw7rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.819371 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.842731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data" (OuterVolumeSpecName: "config-data") pod "aab1dc13-46c4-4776-a5d0-59092fd5ec78" (UID: "aab1dc13-46c4-4776-a5d0-59092fd5ec78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.848376 4754 scope.go:117] "RemoveContainer" containerID="0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.856866 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.856884 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524\": container with ID starting with 0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524 not found: ID does not exist" containerID="0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.856920 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524"} err="failed to get container status \"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524\": rpc error: code = NotFound desc = could not find container \"0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524\": container with ID starting with 0dc5a3943639c01d613151e205f974144444f43dd18725a658d40e3511d5e524 not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.856946 4754 scope.go:117] "RemoveContainer" containerID="272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858285 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858765 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-metadata" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858783 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-metadata" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858797 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-log" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858805 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-log" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858822 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-log" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858829 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-log" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858855 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185f8b44-dd99-48d0-940a-7f14690d6728" containerName="nova-manage" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858865 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="185f8b44-dd99-48d0-940a-7f14690d6728" containerName="nova-manage" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858877 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="dnsmasq-dns" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858883 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="dnsmasq-dns" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858891 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-api" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858897 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-api" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.858910 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="init" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.858916 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="init" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859114 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-log" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859130 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-metadata" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859140 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" containerName="nova-metadata-log" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859149 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" containerName="dnsmasq-dns" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859158 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" containerName="nova-api-api" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.859170 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="185f8b44-dd99-48d0-940a-7f14690d6728" containerName="nova-manage" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.860258 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.861713 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aab1dc13-46c4-4776-a5d0-59092fd5ec78" (UID: "aab1dc13-46c4-4776-a5d0-59092fd5ec78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.862511 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.863877 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b\": container with ID starting with 272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b not found: ID does not exist" containerID="272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.863942 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b"} err="failed to get container status \"272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b\": rpc error: code = NotFound desc = could not find container \"272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b\": container with ID starting with 272cb42ed4f30e88dbfdedf421f7aa4d72ba6a62b27eccdfc0369c2a6f7ded4b not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.863987 4754 scope.go:117] "RemoveContainer" containerID="dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.866930 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "aab1dc13-46c4-4776-a5d0-59092fd5ec78" (UID: "aab1dc13-46c4-4776-a5d0-59092fd5ec78"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.868298 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.902200 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw7rf\" (UniqueName: \"kubernetes.io/projected/aab1dc13-46c4-4776-a5d0-59092fd5ec78-kube-api-access-rw7rf\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.902235 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.902244 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.902252 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab1dc13-46c4-4776-a5d0-59092fd5ec78-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.902261 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1dc13-46c4-4776-a5d0-59092fd5ec78-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.910468 4754 scope.go:117] "RemoveContainer" containerID="b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.944103 4754 scope.go:117] "RemoveContainer" containerID="dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.944694 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414\": container with ID starting with dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414 not found: ID does not exist" containerID="dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.944737 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414"} err="failed to get container status \"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414\": rpc error: code = NotFound desc = could not find container \"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414\": container with ID starting with dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414 not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.944766 4754 scope.go:117] "RemoveContainer" containerID="b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.945118 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245\": container with ID starting with b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245 not found: ID does not exist" containerID="b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945162 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245"} err="failed to get container status \"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245\": rpc error: code = NotFound desc = could not find container \"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245\": container with ID starting with b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245 not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945193 4754 scope.go:117] "RemoveContainer" containerID="dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945472 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414"} err="failed to get container status \"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414\": rpc error: code = NotFound desc = could not find container \"dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414\": container with ID starting with dca9bed0032790c5bcd0fdf95d52d9c05819a0d292d1163dc10e245fba9e6414 not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945492 4754 scope.go:117] "RemoveContainer" containerID="b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945728 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245"} err="failed to get container status \"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245\": rpc error: code = NotFound desc = could not find container \"b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245\": container with ID starting with b316c4d2b1977db3bbb893189d253f9aeab3f8e5549e6a54e4afa97639f06245 not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.945749 4754 scope.go:117] "RemoveContainer" containerID="b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.965474 4754 scope.go:117] "RemoveContainer" containerID="07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.986402 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.990251 4754 scope.go:117] "RemoveContainer" containerID="b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.990532 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b\": container with ID starting with b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b not found: ID does not exist" containerID="b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.990585 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b"} err="failed to get container status \"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b\": rpc error: code = NotFound desc = could not find container \"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b\": container with ID starting with b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.990608 4754 scope.go:117] "RemoveContainer" containerID="07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" Oct 11 03:23:54 crc kubenswrapper[4754]: E1011 03:23:54.990904 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f\": container with ID starting with 07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f not found: ID does not exist" containerID="07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.990989 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f"} err="failed to get container status \"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f\": rpc error: code = NotFound desc = could not find container \"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f\": container with ID starting with 07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.991024 4754 scope.go:117] "RemoveContainer" containerID="b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.991492 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b"} err="failed to get container status \"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b\": rpc error: code = NotFound desc = could not find container \"b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b\": container with ID starting with b07d40f55d346cf419339691f24d583ea910a460fcd509d65c87e407b2c0118b not found: ID does not exist" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.991513 4754 scope.go:117] "RemoveContainer" containerID="07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f" Oct 11 03:23:54 crc kubenswrapper[4754]: I1011 03:23:54.991888 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f"} err="failed to get container status \"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f\": rpc error: code = NotFound desc = could not find container \"07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f\": container with ID starting with 07ae87ae2538263484f7774fc77e51e1cebc50146fbd115ba1a3692d59f1050f not found: ID does not exist" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.004823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.004879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzj8w\" (UniqueName: \"kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.004901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.005127 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.094893 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a5aafb-45cd-47a0-aae1-e564d0d34848" path="/var/lib/kubelet/pods/35a5aafb-45cd-47a0-aae1-e564d0d34848/volumes" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.095848 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f222a5-c392-4e76-a018-856c6c7bd45d" path="/var/lib/kubelet/pods/e0f222a5-c392-4e76-a018-856c6c7bd45d/volumes" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.105896 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle\") pod \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106066 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data\") pod \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wks89\" (UniqueName: \"kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89\") pod \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106134 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts\") pod \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\" (UID: \"a1639fb4-c6c2-48b4-a13c-6f5166229dac\") " Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzj8w\" (UniqueName: \"kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.106579 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.107290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.108893 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts" (OuterVolumeSpecName: "scripts") pod "a1639fb4-c6c2-48b4-a13c-6f5166229dac" (UID: "a1639fb4-c6c2-48b4-a13c-6f5166229dac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.109764 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.110724 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.111264 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89" (OuterVolumeSpecName: "kube-api-access-wks89") pod "a1639fb4-c6c2-48b4-a13c-6f5166229dac" (UID: "a1639fb4-c6c2-48b4-a13c-6f5166229dac"). InnerVolumeSpecName "kube-api-access-wks89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.122216 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzj8w\" (UniqueName: \"kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w\") pod \"nova-api-0\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.131360 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1639fb4-c6c2-48b4-a13c-6f5166229dac" (UID: "a1639fb4-c6c2-48b4-a13c-6f5166229dac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.137434 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data" (OuterVolumeSpecName: "config-data") pod "a1639fb4-c6c2-48b4-a13c-6f5166229dac" (UID: "a1639fb4-c6c2-48b4-a13c-6f5166229dac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.187735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.208517 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.208543 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wks89\" (UniqueName: \"kubernetes.io/projected/a1639fb4-c6c2-48b4-a13c-6f5166229dac-kube-api-access-wks89\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.208552 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.208560 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1639fb4-c6c2-48b4-a13c-6f5166229dac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.617144 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.785774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" event={"ID":"a1639fb4-c6c2-48b4-a13c-6f5166229dac","Type":"ContainerDied","Data":"2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929"} Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.785812 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b37a2680b4ae2fe2a3a460cf2e37f053aeb33264813646c650a9ac9dae0e929" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.785865 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zn2qx" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.810610 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 03:23:55 crc kubenswrapper[4754]: E1011 03:23:55.811075 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1639fb4-c6c2-48b4-a13c-6f5166229dac" containerName="nova-cell1-conductor-db-sync" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.811092 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1639fb4-c6c2-48b4-a13c-6f5166229dac" containerName="nova-cell1-conductor-db-sync" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.811337 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1639fb4-c6c2-48b4-a13c-6f5166229dac" containerName="nova-cell1-conductor-db-sync" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.811909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerStarted","Data":"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d"} Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.811933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerStarted","Data":"e97ddd39ad9175b43a92ade2049370c9f4278d07774e8ee183715d14e7fda447"} Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.812115 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.817004 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.830103 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.921516 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.922586 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6mfj\" (UniqueName: \"kubernetes.io/projected/4401efc4-ecf2-42cd-b406-38206593f37e-kube-api-access-j6mfj\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:55 crc kubenswrapper[4754]: I1011 03:23:55.922640 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.024754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.025248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6mfj\" (UniqueName: \"kubernetes.io/projected/4401efc4-ecf2-42cd-b406-38206593f37e-kube-api-access-j6mfj\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.025310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.029261 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.029611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4401efc4-ecf2-42cd-b406-38206593f37e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.041474 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6mfj\" (UniqueName: \"kubernetes.io/projected/4401efc4-ecf2-42cd-b406-38206593f37e-kube-api-access-j6mfj\") pod \"nova-cell1-conductor-0\" (UID: \"4401efc4-ecf2-42cd-b406-38206593f37e\") " pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.138743 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.496119 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.637735 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.673987 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data\") pod \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.674034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle\") pod \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.674070 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7b8s\" (UniqueName: \"kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s\") pod \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\" (UID: \"1e543740-8abc-41f7-8378-a4d7aaf2ad32\") " Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.678237 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s" (OuterVolumeSpecName: "kube-api-access-s7b8s") pod "1e543740-8abc-41f7-8378-a4d7aaf2ad32" (UID: "1e543740-8abc-41f7-8378-a4d7aaf2ad32"). InnerVolumeSpecName "kube-api-access-s7b8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.695885 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data" (OuterVolumeSpecName: "config-data") pod "1e543740-8abc-41f7-8378-a4d7aaf2ad32" (UID: "1e543740-8abc-41f7-8378-a4d7aaf2ad32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.696292 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e543740-8abc-41f7-8378-a4d7aaf2ad32" (UID: "1e543740-8abc-41f7-8378-a4d7aaf2ad32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.776566 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.776612 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e543740-8abc-41f7-8378-a4d7aaf2ad32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.776626 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7b8s\" (UniqueName: \"kubernetes.io/projected/1e543740-8abc-41f7-8378-a4d7aaf2ad32-kube-api-access-s7b8s\") on node \"crc\" DevicePath \"\"" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.825905 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerStarted","Data":"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b"} Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.831074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4401efc4-ecf2-42cd-b406-38206593f37e","Type":"ContainerStarted","Data":"de6142aafbd3c4cc7317367982015a62dc69344005d2d5f4f1da618f580013e4"} Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.831115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4401efc4-ecf2-42cd-b406-38206593f37e","Type":"ContainerStarted","Data":"9a5c19baac6b7ec7d2f48e214e7bedf833ef7cbf4d8b3332b87fba40a0b5b7e1"} Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.831276 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.835259 4754 generic.go:334] "Generic (PLEG): container finished" podID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" containerID="a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5" exitCode=0 Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.835294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e543740-8abc-41f7-8378-a4d7aaf2ad32","Type":"ContainerDied","Data":"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5"} Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.835317 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e543740-8abc-41f7-8378-a4d7aaf2ad32","Type":"ContainerDied","Data":"66983ef1e858d02a64f64af7372df0a92860a46e02a7806e9a67d24d95836026"} Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.835347 4754 scope.go:117] "RemoveContainer" containerID="a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.835366 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.860161 4754 scope.go:117] "RemoveContainer" containerID="a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5" Oct 11 03:23:56 crc kubenswrapper[4754]: E1011 03:23:56.866672 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5\": container with ID starting with a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5 not found: ID does not exist" containerID="a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.866793 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5"} err="failed to get container status \"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5\": rpc error: code = NotFound desc = could not find container \"a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5\": container with ID starting with a12605d317b6a05a4113dfa12064ffbba12628fa2170103cca2f41b00f30b7f5 not found: ID does not exist" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.872407 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.872373268 podStartE2EDuration="2.872373268s" podCreationTimestamp="2025-10-11 03:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:56.849322897 +0000 UTC m=+1084.408267682" watchObservedRunningTime="2025-10-11 03:23:56.872373268 +0000 UTC m=+1084.431318063" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.907816 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.907785907 podStartE2EDuration="1.907785907s" podCreationTimestamp="2025-10-11 03:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:56.881742332 +0000 UTC m=+1084.440687117" watchObservedRunningTime="2025-10-11 03:23:56.907785907 +0000 UTC m=+1084.466730692" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.931433 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.945487 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.951121 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:56 crc kubenswrapper[4754]: E1011 03:23:56.951510 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" containerName="nova-scheduler-scheduler" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.951537 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" containerName="nova-scheduler-scheduler" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.951747 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" containerName="nova-scheduler-scheduler" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.952421 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.955358 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.961452 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.984230 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.984270 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:56 crc kubenswrapper[4754]: I1011 03:23:56.984296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf5t6\" (UniqueName: \"kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.085599 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.085644 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.085671 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf5t6\" (UniqueName: \"kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.089461 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.089661 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.105370 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf5t6\" (UniqueName: \"kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6\") pod \"nova-scheduler-0\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.108756 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e543740-8abc-41f7-8378-a4d7aaf2ad32" path="/var/lib/kubelet/pods/1e543740-8abc-41f7-8378-a4d7aaf2ad32/volumes" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.268830 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.757285 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:23:57 crc kubenswrapper[4754]: I1011 03:23:57.852129 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6d81c96-0082-4814-9f2c-e486a4bace3c","Type":"ContainerStarted","Data":"03f73ce8a3d56d5404b952127e427b0bc7a051b7d3724e7626ad940a0393ec92"} Oct 11 03:23:58 crc kubenswrapper[4754]: I1011 03:23:58.863702 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6d81c96-0082-4814-9f2c-e486a4bace3c","Type":"ContainerStarted","Data":"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762"} Oct 11 03:23:58 crc kubenswrapper[4754]: I1011 03:23:58.883545 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.883526022 podStartE2EDuration="2.883526022s" podCreationTimestamp="2025-10-11 03:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:23:58.882900105 +0000 UTC m=+1086.441844900" watchObservedRunningTime="2025-10-11 03:23:58.883526022 +0000 UTC m=+1086.442470807" Oct 11 03:24:01 crc kubenswrapper[4754]: I1011 03:24:01.172117 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 11 03:24:02 crc kubenswrapper[4754]: I1011 03:24:02.269295 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 03:24:05 crc kubenswrapper[4754]: I1011 03:24:05.188835 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:05 crc kubenswrapper[4754]: I1011 03:24:05.189598 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:06 crc kubenswrapper[4754]: I1011 03:24:06.229261 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:06 crc kubenswrapper[4754]: I1011 03:24:06.229317 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:07 crc kubenswrapper[4754]: I1011 03:24:07.270006 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 11 03:24:07 crc kubenswrapper[4754]: I1011 03:24:07.302850 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 11 03:24:08 crc kubenswrapper[4754]: I1011 03:24:08.009921 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 11 03:24:15 crc kubenswrapper[4754]: I1011 03:24:15.199502 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 03:24:15 crc kubenswrapper[4754]: I1011 03:24:15.200113 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 03:24:15 crc kubenswrapper[4754]: I1011 03:24:15.204080 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 03:24:15 crc kubenswrapper[4754]: I1011 03:24:15.207383 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.021934 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.026168 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.202660 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.204370 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.217485 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.359759 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.359847 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.359876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.359906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.359928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr7sl\" (UniqueName: \"kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.461611 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.461685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.461707 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.461738 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.461757 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr7sl\" (UniqueName: \"kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.462847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.463689 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.464214 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.464310 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.479052 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr7sl\" (UniqueName: \"kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl\") pod \"dnsmasq-dns-5b856c5697-4lnrj\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:16 crc kubenswrapper[4754]: I1011 03:24:16.538862 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:17 crc kubenswrapper[4754]: I1011 03:24:17.015746 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:24:17 crc kubenswrapper[4754]: I1011 03:24:17.036093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" event={"ID":"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f","Type":"ContainerStarted","Data":"fb8d96504028d4136d4e980c8a02fcd8be4f2cfc5de0edf19833ed50d2d0baa4"} Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.045996 4754 generic.go:334] "Generic (PLEG): container finished" podID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerID="da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d" exitCode=0 Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.046037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" event={"ID":"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f","Type":"ContainerDied","Data":"da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d"} Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.447205 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.447884 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-central-agent" containerID="cri-o://ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3" gracePeriod=30 Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.447946 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="proxy-httpd" containerID="cri-o://94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d" gracePeriod=30 Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.448053 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="sg-core" containerID="cri-o://5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316" gracePeriod=30 Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.448136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-notification-agent" containerID="cri-o://a288bebe112d781f81fad12fbbe9feca13352223f08d90ad5c5c053966be1c91" gracePeriod=30 Oct 11 03:24:18 crc kubenswrapper[4754]: I1011 03:24:18.959083 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058000 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea991e6e-a604-4713-a5a1-fc134a327235" containerID="94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d" exitCode=0 Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058040 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea991e6e-a604-4713-a5a1-fc134a327235" containerID="5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316" exitCode=2 Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058053 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea991e6e-a604-4713-a5a1-fc134a327235" containerID="ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3" exitCode=0 Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058064 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerDied","Data":"94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d"} Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerDied","Data":"5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316"} Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.058138 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerDied","Data":"ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3"} Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.060529 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" event={"ID":"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f","Type":"ContainerStarted","Data":"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f"} Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.060755 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-log" containerID="cri-o://0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d" gracePeriod=30 Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.060836 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.060896 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-api" containerID="cri-o://3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b" gracePeriod=30 Oct 11 03:24:19 crc kubenswrapper[4754]: I1011 03:24:19.100031 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" podStartSLOduration=3.100012382 podStartE2EDuration="3.100012382s" podCreationTimestamp="2025-10-11 03:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:19.095245069 +0000 UTC m=+1106.654189854" watchObservedRunningTime="2025-10-11 03:24:19.100012382 +0000 UTC m=+1106.658957167" Oct 11 03:24:19 crc kubenswrapper[4754]: W1011 03:24:19.709878 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-conmon-0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-conmon-0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d.scope: no such file or directory Oct 11 03:24:19 crc kubenswrapper[4754]: W1011 03:24:19.710166 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d.scope: no such file or directory Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.078989 4754 generic.go:334] "Generic (PLEG): container finished" podID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerID="0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d" exitCode=143 Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.079169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerDied","Data":"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d"} Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.081032 4754 generic.go:334] "Generic (PLEG): container finished" podID="57454bdf-3bd3-444a-938d-f3467c9ec40d" containerID="68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4" exitCode=137 Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.081113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57454bdf-3bd3-444a-938d-f3467c9ec40d","Type":"ContainerDied","Data":"68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4"} Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.081148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57454bdf-3bd3-444a-938d-f3467c9ec40d","Type":"ContainerDied","Data":"0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803"} Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.081163 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.088267 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.163792 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle\") pod \"57454bdf-3bd3-444a-938d-f3467c9ec40d\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.164096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data\") pod \"57454bdf-3bd3-444a-938d-f3467c9ec40d\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.164147 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2tzf\" (UniqueName: \"kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf\") pod \"57454bdf-3bd3-444a-938d-f3467c9ec40d\" (UID: \"57454bdf-3bd3-444a-938d-f3467c9ec40d\") " Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.173161 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf" (OuterVolumeSpecName: "kube-api-access-d2tzf") pod "57454bdf-3bd3-444a-938d-f3467c9ec40d" (UID: "57454bdf-3bd3-444a-938d-f3467c9ec40d"). InnerVolumeSpecName "kube-api-access-d2tzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.202983 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data" (OuterVolumeSpecName: "config-data") pod "57454bdf-3bd3-444a-938d-f3467c9ec40d" (UID: "57454bdf-3bd3-444a-938d-f3467c9ec40d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.226097 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57454bdf-3bd3-444a-938d-f3467c9ec40d" (UID: "57454bdf-3bd3-444a-938d-f3467c9ec40d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.265215 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.265249 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2tzf\" (UniqueName: \"kubernetes.io/projected/57454bdf-3bd3-444a-938d-f3467c9ec40d-kube-api-access-d2tzf\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:20 crc kubenswrapper[4754]: I1011 03:24:20.265260 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57454bdf-3bd3-444a-938d-f3467c9ec40d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.089327 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.137575 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.146609 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.156412 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:24:21 crc kubenswrapper[4754]: E1011 03:24:21.156909 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57454bdf-3bd3-444a-938d-f3467c9ec40d" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.156926 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="57454bdf-3bd3-444a-938d-f3467c9ec40d" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.157181 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="57454bdf-3bd3-444a-938d-f3467c9ec40d" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.158010 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.162239 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.162408 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.162564 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.172624 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.193355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.193517 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzbxf\" (UniqueName: \"kubernetes.io/projected/dd32e65c-d404-481f-aa9a-18e3a0819262-kube-api-access-dzbxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.193611 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.193696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.193742 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.295565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.295643 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzbxf\" (UniqueName: \"kubernetes.io/projected/dd32e65c-d404-481f-aa9a-18e3a0819262-kube-api-access-dzbxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.295691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.295731 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.295758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.301089 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.301208 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.301671 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.313726 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32e65c-d404-481f-aa9a-18e3a0819262-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.316047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzbxf\" (UniqueName: \"kubernetes.io/projected/dd32e65c-d404-481f-aa9a-18e3a0819262-kube-api-access-dzbxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd32e65c-d404-481f-aa9a-18e3a0819262\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:21 crc kubenswrapper[4754]: I1011 03:24:21.504740 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.008703 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.106636 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd32e65c-d404-481f-aa9a-18e3a0819262","Type":"ContainerStarted","Data":"c42de6a425c2fd925fec6aa5a008d0486e16db316917f2bd343b3710b74ee871"} Oct 11 03:24:22 crc kubenswrapper[4754]: E1011 03:24:22.506861 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-conmon-5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57454bdf_3bd3_444a_938d_f3467c9ec40d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-conmon-3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57454bdf_3bd3_444a_938d_f3467c9ec40d.slice/crio-0ceddf6b13ebfacab83f4890b8e28e5038dcaf3fee41d94865722e96c9025803\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57454bdf_3bd3_444a_938d_f3467c9ec40d.slice/crio-conmon-68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-conmon-ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5283f11d_2fef_4403_9d52_6c408371a1ec.slice/crio-3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57454bdf_3bd3_444a_938d_f3467c9ec40d.slice/crio-68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea991e6e_a604_4713_a5a1_fc134a327235.slice/crio-conmon-94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.694419 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.716323 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data\") pod \"5283f11d-2fef-4403-9d52-6c408371a1ec\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.716442 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs\") pod \"5283f11d-2fef-4403-9d52-6c408371a1ec\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.716486 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle\") pod \"5283f11d-2fef-4403-9d52-6c408371a1ec\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.716537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzj8w\" (UniqueName: \"kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w\") pod \"5283f11d-2fef-4403-9d52-6c408371a1ec\" (UID: \"5283f11d-2fef-4403-9d52-6c408371a1ec\") " Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.717561 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs" (OuterVolumeSpecName: "logs") pod "5283f11d-2fef-4403-9d52-6c408371a1ec" (UID: "5283f11d-2fef-4403-9d52-6c408371a1ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.765937 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w" (OuterVolumeSpecName: "kube-api-access-xzj8w") pod "5283f11d-2fef-4403-9d52-6c408371a1ec" (UID: "5283f11d-2fef-4403-9d52-6c408371a1ec"). InnerVolumeSpecName "kube-api-access-xzj8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.776564 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5283f11d-2fef-4403-9d52-6c408371a1ec" (UID: "5283f11d-2fef-4403-9d52-6c408371a1ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.778077 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data" (OuterVolumeSpecName: "config-data") pod "5283f11d-2fef-4403-9d52-6c408371a1ec" (UID: "5283f11d-2fef-4403-9d52-6c408371a1ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.823402 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5283f11d-2fef-4403-9d52-6c408371a1ec-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.823443 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.823457 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzj8w\" (UniqueName: \"kubernetes.io/projected/5283f11d-2fef-4403-9d52-6c408371a1ec-kube-api-access-xzj8w\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:22 crc kubenswrapper[4754]: I1011 03:24:22.823469 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5283f11d-2fef-4403-9d52-6c408371a1ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.096985 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57454bdf-3bd3-444a-938d-f3467c9ec40d" path="/var/lib/kubelet/pods/57454bdf-3bd3-444a-938d-f3467c9ec40d/volumes" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.120322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd32e65c-d404-481f-aa9a-18e3a0819262","Type":"ContainerStarted","Data":"575e47e33889b6b41ea357af4dac6850682f7790f4c6bb82d78c36b99bb91103"} Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.122450 4754 generic.go:334] "Generic (PLEG): container finished" podID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerID="3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b" exitCode=0 Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.122498 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerDied","Data":"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b"} Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.122512 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.122540 4754 scope.go:117] "RemoveContainer" containerID="3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.122528 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5283f11d-2fef-4403-9d52-6c408371a1ec","Type":"ContainerDied","Data":"e97ddd39ad9175b43a92ade2049370c9f4278d07774e8ee183715d14e7fda447"} Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.149219 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.149194534 podStartE2EDuration="2.149194534s" podCreationTimestamp="2025-10-11 03:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:23.134630416 +0000 UTC m=+1110.693575191" watchObservedRunningTime="2025-10-11 03:24:23.149194534 +0000 UTC m=+1110.708139319" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.154932 4754 scope.go:117] "RemoveContainer" containerID="0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.161537 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.180286 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.187772 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:23 crc kubenswrapper[4754]: E1011 03:24:23.188233 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-api" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.188247 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-api" Oct 11 03:24:23 crc kubenswrapper[4754]: E1011 03:24:23.188257 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-log" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.188263 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-log" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.188427 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-log" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.188436 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" containerName="nova-api-api" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.189395 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.192712 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.192932 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.193114 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.194097 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.211795 4754 scope.go:117] "RemoveContainer" containerID="3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b" Oct 11 03:24:23 crc kubenswrapper[4754]: E1011 03:24:23.212328 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b\": container with ID starting with 3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b not found: ID does not exist" containerID="3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.212364 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b"} err="failed to get container status \"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b\": rpc error: code = NotFound desc = could not find container \"3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b\": container with ID starting with 3b29e5d046888592e08d22ec4092cf92bb704e2924c600ae76777cc47afe171b not found: ID does not exist" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.212390 4754 scope.go:117] "RemoveContainer" containerID="0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d" Oct 11 03:24:23 crc kubenswrapper[4754]: E1011 03:24:23.212772 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d\": container with ID starting with 0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d not found: ID does not exist" containerID="0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.212793 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d"} err="failed to get container status \"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d\": rpc error: code = NotFound desc = could not find container \"0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d\": container with ID starting with 0a3e00efa876f07be8628b7663c246a28317472cf72fd77815b95e613aae1d0d not found: ID does not exist" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233288 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wblw5\" (UniqueName: \"kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233357 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233416 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233438 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233451 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.233470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335346 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335405 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335512 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wblw5\" (UniqueName: \"kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335551 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.335921 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.340045 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.342089 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.342572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.344409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.357040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wblw5\" (UniqueName: \"kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5\") pod \"nova-api-0\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " pod="openstack/nova-api-0" Oct 11 03:24:23 crc kubenswrapper[4754]: I1011 03:24:23.513460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:24 crc kubenswrapper[4754]: I1011 03:24:24.022242 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:24 crc kubenswrapper[4754]: I1011 03:24:24.133768 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerStarted","Data":"3269f134554f664446c022caddb2b34dcd22218543d292fd39abaea8dfb9c06b"} Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.099748 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5283f11d-2fef-4403-9d52-6c408371a1ec" path="/var/lib/kubelet/pods/5283f11d-2fef-4403-9d52-6c408371a1ec/volumes" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.149110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerStarted","Data":"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f"} Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.149187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerStarted","Data":"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a"} Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.174690 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.174672269 podStartE2EDuration="2.174672269s" podCreationTimestamp="2025-10-11 03:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:25.169673339 +0000 UTC m=+1112.728618124" watchObservedRunningTime="2025-10-11 03:24:25.174672269 +0000 UTC m=+1112.733617054" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.208748 4754 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podaab1dc13-46c4-4776-a5d0-59092fd5ec78"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podaab1dc13-46c4-4776-a5d0-59092fd5ec78] : Timed out while waiting for systemd to remove kubepods-besteffort-podaab1dc13_46c4_4776_a5d0_59092fd5ec78.slice" Oct 11 03:24:25 crc kubenswrapper[4754]: E1011 03:24:25.208861 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podaab1dc13-46c4-4776-a5d0-59092fd5ec78] : unable to destroy cgroup paths for cgroup [kubepods besteffort podaab1dc13-46c4-4776-a5d0-59092fd5ec78] : Timed out while waiting for systemd to remove kubepods-besteffort-podaab1dc13_46c4_4776_a5d0_59092fd5ec78.slice" pod="openstack/nova-metadata-0" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.221461 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea991e6e-a604-4713-a5a1-fc134a327235" containerID="a288bebe112d781f81fad12fbbe9feca13352223f08d90ad5c5c053966be1c91" exitCode=0 Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.221533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerDied","Data":"a288bebe112d781f81fad12fbbe9feca13352223f08d90ad5c5c053966be1c91"} Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.522380 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615595 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615738 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615758 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615800 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615816 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzwt9\" (UniqueName: \"kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.615986 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd\") pod \"ea991e6e-a604-4713-a5a1-fc134a327235\" (UID: \"ea991e6e-a604-4713-a5a1-fc134a327235\") " Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.616877 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.617356 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.623189 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts" (OuterVolumeSpecName: "scripts") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.625609 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9" (OuterVolumeSpecName: "kube-api-access-xzwt9") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "kube-api-access-xzwt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.649030 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.695047 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.718816 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.719084 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.719149 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.719204 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.719358 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea991e6e-a604-4713-a5a1-fc134a327235-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.719414 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzwt9\" (UniqueName: \"kubernetes.io/projected/ea991e6e-a604-4713-a5a1-fc134a327235-kube-api-access-xzwt9\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.735934 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data" (OuterVolumeSpecName: "config-data") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.737678 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea991e6e-a604-4713-a5a1-fc134a327235" (UID: "ea991e6e-a604-4713-a5a1-fc134a327235"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.822201 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:25 crc kubenswrapper[4754]: I1011 03:24:25.822259 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea991e6e-a604-4713-a5a1-fc134a327235-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.233148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea991e6e-a604-4713-a5a1-fc134a327235","Type":"ContainerDied","Data":"2ed8321b57c4b4bd505a5db05b5ffdc581e89ffdb6a8090ac638e743cb0ed658"} Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.233177 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.233500 4754 scope.go:117] "RemoveContainer" containerID="94f59b47a99589d6994cf70dd6ded03ff8db81898cf9a9d6669ac7dd4125a02d" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.233250 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.256148 4754 scope.go:117] "RemoveContainer" containerID="5375d67b5d615930ecd7c442ea9da73e532dd87e554e30b34a721306df7dd316" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.300369 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.311749 4754 scope.go:117] "RemoveContainer" containerID="a288bebe112d781f81fad12fbbe9feca13352223f08d90ad5c5c053966be1c91" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.316238 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.331828 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.351956 4754 scope.go:117] "RemoveContainer" containerID="ed102c7639f6ad560f7b25426eeea1634e9aaffc3693691b4b234fbd92a781c3" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.395164 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.402651 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: E1011 03:24:26.403051 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="proxy-httpd" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403069 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="proxy-httpd" Oct 11 03:24:26 crc kubenswrapper[4754]: E1011 03:24:26.403086 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-notification-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403092 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-notification-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: E1011 03:24:26.403103 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="sg-core" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403110 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="sg-core" Oct 11 03:24:26 crc kubenswrapper[4754]: E1011 03:24:26.403119 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-central-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403126 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-central-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403352 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-central-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403364 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="ceilometer-notification-agent" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403377 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="sg-core" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.403388 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" containerName="proxy-httpd" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.405066 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.409158 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.409353 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.412796 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.428804 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.430451 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.434185 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.434495 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.438018 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.445691 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.505307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536042 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmkbk\" (UniqueName: \"kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536122 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536180 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536250 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536275 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536318 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp75w\" (UniqueName: \"kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536365 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536407 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536507 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.536652 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.541283 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.589442 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.589662 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="dnsmasq-dns" containerID="cri-o://35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed" gracePeriod=10 Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639220 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmkbk\" (UniqueName: \"kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp75w\" (UniqueName: \"kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.639621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.647091 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.647819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.648386 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.649349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.652497 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.652829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.653121 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.653487 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.654232 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.654512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.657101 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.665084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp75w\" (UniqueName: \"kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w\") pod \"nova-metadata-0\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " pod="openstack/nova-metadata-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.673715 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmkbk\" (UniqueName: \"kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk\") pod \"ceilometer-0\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.761862 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:24:26 crc kubenswrapper[4754]: I1011 03:24:26.768622 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.022998 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.110862 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab1dc13-46c4-4776-a5d0-59092fd5ec78" path="/var/lib/kubelet/pods/aab1dc13-46c4-4776-a5d0-59092fd5ec78/volumes" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.111624 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea991e6e-a604-4713-a5a1-fc134a327235" path="/var/lib/kubelet/pods/ea991e6e-a604-4713-a5a1-fc134a327235/volumes" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.148855 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config\") pod \"ab70a33e-1420-4d1d-9af7-b2f93e767536\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.148939 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw2q2\" (UniqueName: \"kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2\") pod \"ab70a33e-1420-4d1d-9af7-b2f93e767536\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.149052 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb\") pod \"ab70a33e-1420-4d1d-9af7-b2f93e767536\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.149166 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc\") pod \"ab70a33e-1420-4d1d-9af7-b2f93e767536\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.149279 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb\") pod \"ab70a33e-1420-4d1d-9af7-b2f93e767536\" (UID: \"ab70a33e-1420-4d1d-9af7-b2f93e767536\") " Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.154706 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2" (OuterVolumeSpecName: "kube-api-access-gw2q2") pod "ab70a33e-1420-4d1d-9af7-b2f93e767536" (UID: "ab70a33e-1420-4d1d-9af7-b2f93e767536"). InnerVolumeSpecName "kube-api-access-gw2q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.197120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config" (OuterVolumeSpecName: "config") pod "ab70a33e-1420-4d1d-9af7-b2f93e767536" (UID: "ab70a33e-1420-4d1d-9af7-b2f93e767536"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.198547 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab70a33e-1420-4d1d-9af7-b2f93e767536" (UID: "ab70a33e-1420-4d1d-9af7-b2f93e767536"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.202272 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab70a33e-1420-4d1d-9af7-b2f93e767536" (UID: "ab70a33e-1420-4d1d-9af7-b2f93e767536"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.208428 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab70a33e-1420-4d1d-9af7-b2f93e767536" (UID: "ab70a33e-1420-4d1d-9af7-b2f93e767536"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.243438 4754 generic.go:334] "Generic (PLEG): container finished" podID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerID="35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed" exitCode=0 Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.243492 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" event={"ID":"ab70a33e-1420-4d1d-9af7-b2f93e767536","Type":"ContainerDied","Data":"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed"} Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.243527 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" event={"ID":"ab70a33e-1420-4d1d-9af7-b2f93e767536","Type":"ContainerDied","Data":"c5b28be5461fd518d215e9ffb4037b64a24b53d87a5b1c2fd0a21075e77a1488"} Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.243523 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-l6xgx" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.243602 4754 scope.go:117] "RemoveContainer" containerID="35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.250839 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.251038 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.251050 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw2q2\" (UniqueName: \"kubernetes.io/projected/ab70a33e-1420-4d1d-9af7-b2f93e767536-kube-api-access-gw2q2\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.251060 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.251090 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab70a33e-1420-4d1d-9af7-b2f93e767536-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.277150 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.284644 4754 scope.go:117] "RemoveContainer" containerID="d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.285078 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-l6xgx"] Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.293513 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.312355 4754 scope.go:117] "RemoveContainer" containerID="35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed" Oct 11 03:24:27 crc kubenswrapper[4754]: E1011 03:24:27.316732 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed\": container with ID starting with 35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed not found: ID does not exist" containerID="35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.316796 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed"} err="failed to get container status \"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed\": rpc error: code = NotFound desc = could not find container \"35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed\": container with ID starting with 35ce23848cdf42974c596a6755382bfb777c42fd66dd8a6da1431ace10f645ed not found: ID does not exist" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.316833 4754 scope.go:117] "RemoveContainer" containerID="d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d" Oct 11 03:24:27 crc kubenswrapper[4754]: E1011 03:24:27.317771 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d\": container with ID starting with d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d not found: ID does not exist" containerID="d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.317813 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d"} err="failed to get container status \"d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d\": rpc error: code = NotFound desc = could not find container \"d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d\": container with ID starting with d182242343398eb46271e62cb77c03e856b65a36f63638e52a14ba4affe4f94d not found: ID does not exist" Oct 11 03:24:27 crc kubenswrapper[4754]: I1011 03:24:27.439321 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:27 crc kubenswrapper[4754]: W1011 03:24:27.449109 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc39d8ae6_3eb4_4bec_9800_709939be7dc5.slice/crio-d38f474ba1fde127f8e929326211e55d985862af6035106f008da70a7f912eab WatchSource:0}: Error finding container d38f474ba1fde127f8e929326211e55d985862af6035106f008da70a7f912eab: Status 404 returned error can't find the container with id d38f474ba1fde127f8e929326211e55d985862af6035106f008da70a7f912eab Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.252329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerStarted","Data":"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf"} Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.252732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerStarted","Data":"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565"} Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.252744 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerStarted","Data":"d38f474ba1fde127f8e929326211e55d985862af6035106f008da70a7f912eab"} Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.253700 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerStarted","Data":"f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592"} Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.253737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerStarted","Data":"6a8f61075043d247e92311c189fa15b632341c512b4a5b65e9cd09e4bb1b7d7c"} Oct 11 03:24:28 crc kubenswrapper[4754]: I1011 03:24:28.270547 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.270532344 podStartE2EDuration="2.270532344s" podCreationTimestamp="2025-10-11 03:24:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:28.268809335 +0000 UTC m=+1115.827754120" watchObservedRunningTime="2025-10-11 03:24:28.270532344 +0000 UTC m=+1115.829477129" Oct 11 03:24:29 crc kubenswrapper[4754]: I1011 03:24:29.096276 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" path="/var/lib/kubelet/pods/ab70a33e-1420-4d1d-9af7-b2f93e767536/volumes" Oct 11 03:24:29 crc kubenswrapper[4754]: I1011 03:24:29.264031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerStarted","Data":"38127e3f1de4bafc2552a6c1b37bf8bdfb882b7b76734b132471363d941e2efb"} Oct 11 03:24:30 crc kubenswrapper[4754]: I1011 03:24:30.274349 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerStarted","Data":"669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903"} Oct 11 03:24:31 crc kubenswrapper[4754]: I1011 03:24:31.505291 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:31 crc kubenswrapper[4754]: I1011 03:24:31.526194 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:31 crc kubenswrapper[4754]: I1011 03:24:31.769493 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:24:31 crc kubenswrapper[4754]: I1011 03:24:31.770211 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.301778 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerStarted","Data":"27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06"} Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.321816 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.488022902 podStartE2EDuration="6.321799284s" podCreationTimestamp="2025-10-11 03:24:26 +0000 UTC" firstStartedPulling="2025-10-11 03:24:27.312259088 +0000 UTC m=+1114.871203873" lastFinishedPulling="2025-10-11 03:24:31.14603545 +0000 UTC m=+1118.704980255" observedRunningTime="2025-10-11 03:24:32.319270833 +0000 UTC m=+1119.878215608" watchObservedRunningTime="2025-10-11 03:24:32.321799284 +0000 UTC m=+1119.880744069" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.325624 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.513145 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mfrtj"] Oct 11 03:24:32 crc kubenswrapper[4754]: E1011 03:24:32.514637 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="dnsmasq-dns" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.514725 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="dnsmasq-dns" Oct 11 03:24:32 crc kubenswrapper[4754]: E1011 03:24:32.514793 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="init" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.514848 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="init" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.515127 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab70a33e-1420-4d1d-9af7-b2f93e767536" containerName="dnsmasq-dns" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.515795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.518117 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.519159 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.526893 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mfrtj"] Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.570549 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.570610 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.570840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.571013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9nkt\" (UniqueName: \"kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.672443 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.672539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.672787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.672927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9nkt\" (UniqueName: \"kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.680888 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.682173 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.685783 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.688436 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9nkt\" (UniqueName: \"kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt\") pod \"nova-cell1-cell-mapping-mfrtj\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:32 crc kubenswrapper[4754]: I1011 03:24:32.837452 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:33 crc kubenswrapper[4754]: I1011 03:24:33.310283 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:24:33 crc kubenswrapper[4754]: I1011 03:24:33.320274 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mfrtj"] Oct 11 03:24:33 crc kubenswrapper[4754]: I1011 03:24:33.513892 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:33 crc kubenswrapper[4754]: I1011 03:24:33.514282 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:34 crc kubenswrapper[4754]: I1011 03:24:34.324592 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mfrtj" event={"ID":"ab4c309f-66cb-46e5-b6b0-8599a4019cd3","Type":"ContainerStarted","Data":"32fe382470935e34e03dadf3080fb4856f160b0a457434596b165c1adf7d06c7"} Oct 11 03:24:34 crc kubenswrapper[4754]: I1011 03:24:34.324662 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mfrtj" event={"ID":"ab4c309f-66cb-46e5-b6b0-8599a4019cd3","Type":"ContainerStarted","Data":"79d91a769e8cf4725961ba8e08f83e4f4b960a4caa8d0cf032d5b39a49083476"} Oct 11 03:24:34 crc kubenswrapper[4754]: I1011 03:24:34.343662 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mfrtj" podStartSLOduration=2.343640497 podStartE2EDuration="2.343640497s" podCreationTimestamp="2025-10-11 03:24:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:34.338584046 +0000 UTC m=+1121.897528841" watchObservedRunningTime="2025-10-11 03:24:34.343640497 +0000 UTC m=+1121.902585302" Oct 11 03:24:34 crc kubenswrapper[4754]: I1011 03:24:34.528091 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.182:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:34 crc kubenswrapper[4754]: I1011 03:24:34.528091 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.182:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:36 crc kubenswrapper[4754]: I1011 03:24:36.768829 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 03:24:36 crc kubenswrapper[4754]: I1011 03:24:36.770742 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 03:24:37 crc kubenswrapper[4754]: I1011 03:24:37.779212 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:37 crc kubenswrapper[4754]: I1011 03:24:37.779310 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:38 crc kubenswrapper[4754]: I1011 03:24:38.366128 4754 generic.go:334] "Generic (PLEG): container finished" podID="ab4c309f-66cb-46e5-b6b0-8599a4019cd3" containerID="32fe382470935e34e03dadf3080fb4856f160b0a457434596b165c1adf7d06c7" exitCode=0 Oct 11 03:24:38 crc kubenswrapper[4754]: I1011 03:24:38.366190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mfrtj" event={"ID":"ab4c309f-66cb-46e5-b6b0-8599a4019cd3","Type":"ContainerDied","Data":"32fe382470935e34e03dadf3080fb4856f160b0a457434596b165c1adf7d06c7"} Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.679556 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.710645 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9nkt\" (UniqueName: \"kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt\") pod \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.711011 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts\") pod \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.711165 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle\") pod \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.711291 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data\") pod \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\" (UID: \"ab4c309f-66cb-46e5-b6b0-8599a4019cd3\") " Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.717310 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt" (OuterVolumeSpecName: "kube-api-access-v9nkt") pod "ab4c309f-66cb-46e5-b6b0-8599a4019cd3" (UID: "ab4c309f-66cb-46e5-b6b0-8599a4019cd3"). InnerVolumeSpecName "kube-api-access-v9nkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.718338 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts" (OuterVolumeSpecName: "scripts") pod "ab4c309f-66cb-46e5-b6b0-8599a4019cd3" (UID: "ab4c309f-66cb-46e5-b6b0-8599a4019cd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.742386 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data" (OuterVolumeSpecName: "config-data") pod "ab4c309f-66cb-46e5-b6b0-8599a4019cd3" (UID: "ab4c309f-66cb-46e5-b6b0-8599a4019cd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.743703 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab4c309f-66cb-46e5-b6b0-8599a4019cd3" (UID: "ab4c309f-66cb-46e5-b6b0-8599a4019cd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.813997 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.814052 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9nkt\" (UniqueName: \"kubernetes.io/projected/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-kube-api-access-v9nkt\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.814072 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:39 crc kubenswrapper[4754]: I1011 03:24:39.814087 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c309f-66cb-46e5-b6b0-8599a4019cd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.384600 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mfrtj" event={"ID":"ab4c309f-66cb-46e5-b6b0-8599a4019cd3","Type":"ContainerDied","Data":"79d91a769e8cf4725961ba8e08f83e4f4b960a4caa8d0cf032d5b39a49083476"} Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.384645 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79d91a769e8cf4725961ba8e08f83e4f4b960a4caa8d0cf032d5b39a49083476" Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.385154 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mfrtj" Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.577823 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.578452 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerName="nova-scheduler-scheduler" containerID="cri-o://361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" gracePeriod=30 Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.590180 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.590470 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-log" containerID="cri-o://e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a" gracePeriod=30 Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.590581 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-api" containerID="cri-o://883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f" gracePeriod=30 Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.608577 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.609092 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-log" containerID="cri-o://dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565" gracePeriod=30 Oct 11 03:24:40 crc kubenswrapper[4754]: I1011 03:24:40.609167 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-metadata" containerID="cri-o://8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf" gracePeriod=30 Oct 11 03:24:41 crc kubenswrapper[4754]: I1011 03:24:41.396610 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerID="e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a" exitCode=143 Oct 11 03:24:41 crc kubenswrapper[4754]: I1011 03:24:41.396679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerDied","Data":"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a"} Oct 11 03:24:41 crc kubenswrapper[4754]: I1011 03:24:41.400531 4754 generic.go:334] "Generic (PLEG): container finished" podID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerID="dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565" exitCode=143 Oct 11 03:24:41 crc kubenswrapper[4754]: I1011 03:24:41.400564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerDied","Data":"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565"} Oct 11 03:24:42 crc kubenswrapper[4754]: E1011 03:24:42.271252 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 11 03:24:42 crc kubenswrapper[4754]: E1011 03:24:42.272647 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 11 03:24:42 crc kubenswrapper[4754]: E1011 03:24:42.274571 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 11 03:24:42 crc kubenswrapper[4754]: E1011 03:24:42.274653 4754 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerName="nova-scheduler-scheduler" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.240564 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.247672 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295452 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp75w\" (UniqueName: \"kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w\") pod \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295519 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295547 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295579 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs\") pod \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295691 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle\") pod \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295854 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.295911 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.296060 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.296096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wblw5\" (UniqueName: \"kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5\") pod \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\" (UID: \"0a744dec-454f-43a8-8505-5ca3b6ce5c49\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.296151 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data\") pod \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.296214 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs\") pod \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\" (UID: \"c39d8ae6-3eb4-4bec-9800-709939be7dc5\") " Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.297198 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs" (OuterVolumeSpecName: "logs") pod "c39d8ae6-3eb4-4bec-9800-709939be7dc5" (UID: "c39d8ae6-3eb4-4bec-9800-709939be7dc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.299452 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs" (OuterVolumeSpecName: "logs") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.312192 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5" (OuterVolumeSpecName: "kube-api-access-wblw5") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "kube-api-access-wblw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.312243 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w" (OuterVolumeSpecName: "kube-api-access-lp75w") pod "c39d8ae6-3eb4-4bec-9800-709939be7dc5" (UID: "c39d8ae6-3eb4-4bec-9800-709939be7dc5"). InnerVolumeSpecName "kube-api-access-lp75w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.324298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data" (OuterVolumeSpecName: "config-data") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.327017 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c39d8ae6-3eb4-4bec-9800-709939be7dc5" (UID: "c39d8ae6-3eb4-4bec-9800-709939be7dc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.328658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data" (OuterVolumeSpecName: "config-data") pod "c39d8ae6-3eb4-4bec-9800-709939be7dc5" (UID: "c39d8ae6-3eb4-4bec-9800-709939be7dc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.343705 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c39d8ae6-3eb4-4bec-9800-709939be7dc5" (UID: "c39d8ae6-3eb4-4bec-9800-709939be7dc5"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.345918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.348799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.349513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a744dec-454f-43a8-8505-5ca3b6ce5c49" (UID: "0a744dec-454f-43a8-8505-5ca3b6ce5c49"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398452 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398483 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398493 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398503 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398512 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398521 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a744dec-454f-43a8-8505-5ca3b6ce5c49-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398530 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a744dec-454f-43a8-8505-5ca3b6ce5c49-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398538 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wblw5\" (UniqueName: \"kubernetes.io/projected/0a744dec-454f-43a8-8505-5ca3b6ce5c49-kube-api-access-wblw5\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398548 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c39d8ae6-3eb4-4bec-9800-709939be7dc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398555 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c39d8ae6-3eb4-4bec-9800-709939be7dc5-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.398563 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp75w\" (UniqueName: \"kubernetes.io/projected/c39d8ae6-3eb4-4bec-9800-709939be7dc5-kube-api-access-lp75w\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.426813 4754 generic.go:334] "Generic (PLEG): container finished" podID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerID="8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf" exitCode=0 Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.426875 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.426907 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerDied","Data":"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf"} Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.426993 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c39d8ae6-3eb4-4bec-9800-709939be7dc5","Type":"ContainerDied","Data":"d38f474ba1fde127f8e929326211e55d985862af6035106f008da70a7f912eab"} Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.427019 4754 scope.go:117] "RemoveContainer" containerID="8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.430560 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerID="883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f" exitCode=0 Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.430598 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerDied","Data":"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f"} Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.430620 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a744dec-454f-43a8-8505-5ca3b6ce5c49","Type":"ContainerDied","Data":"3269f134554f664446c022caddb2b34dcd22218543d292fd39abaea8dfb9c06b"} Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.430683 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.461081 4754 scope.go:117] "RemoveContainer" containerID="dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.470023 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.497249 4754 scope.go:117] "RemoveContainer" containerID="8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.497813 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.499342 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf\": container with ID starting with 8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf not found: ID does not exist" containerID="8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.499396 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf"} err="failed to get container status \"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf\": rpc error: code = NotFound desc = could not find container \"8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf\": container with ID starting with 8998c2d6df649a8f72556a3601425e3d991c715f14ff883277fa6aece3f2d8cf not found: ID does not exist" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.499443 4754 scope.go:117] "RemoveContainer" containerID="dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.499837 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565\": container with ID starting with dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565 not found: ID does not exist" containerID="dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.499872 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565"} err="failed to get container status \"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565\": rpc error: code = NotFound desc = could not find container \"dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565\": container with ID starting with dc596c249c4619b6fff8bc271f6e90c99ae5e568ea7c461397ec5d33b9a14565 not found: ID does not exist" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.499899 4754 scope.go:117] "RemoveContainer" containerID="883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.506438 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.518004 4754 scope.go:117] "RemoveContainer" containerID="e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.523073 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.534258 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.534845 4754 scope.go:117] "RemoveContainer" containerID="883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535094 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-log" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535116 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-log" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535127 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-metadata" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535133 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-metadata" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535150 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c309f-66cb-46e5-b6b0-8599a4019cd3" containerName="nova-manage" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535156 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c309f-66cb-46e5-b6b0-8599a4019cd3" containerName="nova-manage" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535169 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-api" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535175 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-api" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535186 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-log" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535191 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-log" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535361 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-api" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535379 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" containerName="nova-api-log" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535386 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-log" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535394 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" containerName="nova-metadata-metadata" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535402 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4c309f-66cb-46e5-b6b0-8599a4019cd3" containerName="nova-manage" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.535886 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f\": container with ID starting with 883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f not found: ID does not exist" containerID="883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535938 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f"} err="failed to get container status \"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f\": rpc error: code = NotFound desc = could not find container \"883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f\": container with ID starting with 883d4f37733cff959c75b8b2c9fe773de55951bef2194c998a1bee5e1e91106f not found: ID does not exist" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.535976 4754 scope.go:117] "RemoveContainer" containerID="e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a" Oct 11 03:24:44 crc kubenswrapper[4754]: E1011 03:24:44.536295 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a\": container with ID starting with e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a not found: ID does not exist" containerID="e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.536324 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a"} err="failed to get container status \"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a\": rpc error: code = NotFound desc = could not find container \"e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a\": container with ID starting with e489782e70b98cebf5558b5781cf71ea81e3445fd7c939b15c6eb311ae9b2c2a not found: ID does not exist" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.536524 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.539554 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.539668 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.543274 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.545284 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.548348 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.548602 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.549192 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.551386 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.560118 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw52z\" (UniqueName: \"kubernetes.io/projected/cd491934-e354-499d-ad2d-f77049e8e4b7-kube-api-access-lw52z\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602135 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd491934-e354-499d-ad2d-f77049e8e4b7-logs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602189 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-logs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602260 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw6xt\" (UniqueName: \"kubernetes.io/projected/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-kube-api-access-hw6xt\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602285 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-public-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602524 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602569 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-config-data\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602610 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602639 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-config-data\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.602670 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704315 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704366 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw6xt\" (UniqueName: \"kubernetes.io/projected/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-kube-api-access-hw6xt\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-public-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704453 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704469 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-config-data\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704490 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704566 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-config-data\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704593 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw52z\" (UniqueName: \"kubernetes.io/projected/cd491934-e354-499d-ad2d-f77049e8e4b7-kube-api-access-lw52z\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704650 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd491934-e354-499d-ad2d-f77049e8e4b7-logs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704678 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-logs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.704989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-logs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.705712 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd491934-e354-499d-ad2d-f77049e8e4b7-logs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.709024 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-config-data\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.709088 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.709106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.709446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-public-tls-certs\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.709886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-config-data\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.711587 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.711681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd491934-e354-499d-ad2d-f77049e8e4b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.720433 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw6xt\" (UniqueName: \"kubernetes.io/projected/21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b-kube-api-access-hw6xt\") pod \"nova-api-0\" (UID: \"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b\") " pod="openstack/nova-api-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.723070 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw52z\" (UniqueName: \"kubernetes.io/projected/cd491934-e354-499d-ad2d-f77049e8e4b7-kube-api-access-lw52z\") pod \"nova-metadata-0\" (UID: \"cd491934-e354-499d-ad2d-f77049e8e4b7\") " pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.858525 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 03:24:44 crc kubenswrapper[4754]: I1011 03:24:44.864600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.111439 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a744dec-454f-43a8-8505-5ca3b6ce5c49" path="/var/lib/kubelet/pods/0a744dec-454f-43a8-8505-5ca3b6ce5c49/volumes" Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.112723 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39d8ae6-3eb4-4bec-9800-709939be7dc5" path="/var/lib/kubelet/pods/c39d8ae6-3eb4-4bec-9800-709939be7dc5/volumes" Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.305823 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.367739 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 03:24:45 crc kubenswrapper[4754]: W1011 03:24:45.388988 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd491934_e354_499d_ad2d_f77049e8e4b7.slice/crio-d4c9985a0bba566c6bf36b52892940620677e6db3db282ae6bc4a18ed00e41e0 WatchSource:0}: Error finding container d4c9985a0bba566c6bf36b52892940620677e6db3db282ae6bc4a18ed00e41e0: Status 404 returned error can't find the container with id d4c9985a0bba566c6bf36b52892940620677e6db3db282ae6bc4a18ed00e41e0 Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.442533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b","Type":"ContainerStarted","Data":"da98a64bfca294cfaf8c0f1d91985f127b8778174e7e0aa9a0fa00458d44cb81"} Oct 11 03:24:45 crc kubenswrapper[4754]: I1011 03:24:45.445719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd491934-e354-499d-ad2d-f77049e8e4b7","Type":"ContainerStarted","Data":"d4c9985a0bba566c6bf36b52892940620677e6db3db282ae6bc4a18ed00e41e0"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.031182 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.128818 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf5t6\" (UniqueName: \"kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6\") pod \"f6d81c96-0082-4814-9f2c-e486a4bace3c\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.128914 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data\") pod \"f6d81c96-0082-4814-9f2c-e486a4bace3c\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.129053 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle\") pod \"f6d81c96-0082-4814-9f2c-e486a4bace3c\" (UID: \"f6d81c96-0082-4814-9f2c-e486a4bace3c\") " Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.132789 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6" (OuterVolumeSpecName: "kube-api-access-zf5t6") pod "f6d81c96-0082-4814-9f2c-e486a4bace3c" (UID: "f6d81c96-0082-4814-9f2c-e486a4bace3c"). InnerVolumeSpecName "kube-api-access-zf5t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.157802 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data" (OuterVolumeSpecName: "config-data") pod "f6d81c96-0082-4814-9f2c-e486a4bace3c" (UID: "f6d81c96-0082-4814-9f2c-e486a4bace3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.160271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6d81c96-0082-4814-9f2c-e486a4bace3c" (UID: "f6d81c96-0082-4814-9f2c-e486a4bace3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.230731 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.230762 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d81c96-0082-4814-9f2c-e486a4bace3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.230772 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf5t6\" (UniqueName: \"kubernetes.io/projected/f6d81c96-0082-4814-9f2c-e486a4bace3c-kube-api-access-zf5t6\") on node \"crc\" DevicePath \"\"" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.458601 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b","Type":"ContainerStarted","Data":"5ca9c109c59449acf0f383e27faacf8a18876ad83769d840d90162d83173b18d"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.458659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b","Type":"ContainerStarted","Data":"36c6ac98d74a0fcc9dd249931239480ed4ed0a7426ad556c83484e001ae033b7"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.461203 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" exitCode=0 Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.461242 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.461292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6d81c96-0082-4814-9f2c-e486a4bace3c","Type":"ContainerDied","Data":"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.461339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6d81c96-0082-4814-9f2c-e486a4bace3c","Type":"ContainerDied","Data":"03f73ce8a3d56d5404b952127e427b0bc7a051b7d3724e7626ad940a0393ec92"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.461356 4754 scope.go:117] "RemoveContainer" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.464497 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd491934-e354-499d-ad2d-f77049e8e4b7","Type":"ContainerStarted","Data":"f8f801d5f650810b56d648c5858c7d29e7f1bc619982b2a3b4ec30b77f310479"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.464532 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd491934-e354-499d-ad2d-f77049e8e4b7","Type":"ContainerStarted","Data":"421cbc594c40d6ac70b5cd5d1a35ba240b4dd9354355df360755d0e3b4628672"} Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.478822 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.478801127 podStartE2EDuration="2.478801127s" podCreationTimestamp="2025-10-11 03:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:46.472988824 +0000 UTC m=+1134.031933609" watchObservedRunningTime="2025-10-11 03:24:46.478801127 +0000 UTC m=+1134.037745902" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.493614 4754 scope.go:117] "RemoveContainer" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" Oct 11 03:24:46 crc kubenswrapper[4754]: E1011 03:24:46.494115 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762\": container with ID starting with 361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762 not found: ID does not exist" containerID="361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.494140 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762"} err="failed to get container status \"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762\": rpc error: code = NotFound desc = could not find container \"361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762\": container with ID starting with 361ae0d1ea62f074af9714be2574f16a3f1e0ad29770e43e22305b5c78470762 not found: ID does not exist" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.502759 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.502739778 podStartE2EDuration="2.502739778s" podCreationTimestamp="2025-10-11 03:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:46.501318719 +0000 UTC m=+1134.060263514" watchObservedRunningTime="2025-10-11 03:24:46.502739778 +0000 UTC m=+1134.061684573" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.525255 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.537186 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.555754 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:46 crc kubenswrapper[4754]: E1011 03:24:46.556292 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerName="nova-scheduler-scheduler" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.556318 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerName="nova-scheduler-scheduler" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.556545 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" containerName="nova-scheduler-scheduler" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.557486 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.559809 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.564874 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.637290 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.637408 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29bg8\" (UniqueName: \"kubernetes.io/projected/d6f98545-924e-4087-8f8e-15c9f9b260cb-kube-api-access-29bg8\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.637596 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-config-data\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.739865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.740029 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29bg8\" (UniqueName: \"kubernetes.io/projected/d6f98545-924e-4087-8f8e-15c9f9b260cb-kube-api-access-29bg8\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.740105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-config-data\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.744888 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.745519 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f98545-924e-4087-8f8e-15c9f9b260cb-config-data\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.757407 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29bg8\" (UniqueName: \"kubernetes.io/projected/d6f98545-924e-4087-8f8e-15c9f9b260cb-kube-api-access-29bg8\") pod \"nova-scheduler-0\" (UID: \"d6f98545-924e-4087-8f8e-15c9f9b260cb\") " pod="openstack/nova-scheduler-0" Oct 11 03:24:46 crc kubenswrapper[4754]: I1011 03:24:46.880015 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 03:24:47 crc kubenswrapper[4754]: I1011 03:24:47.095760 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d81c96-0082-4814-9f2c-e486a4bace3c" path="/var/lib/kubelet/pods/f6d81c96-0082-4814-9f2c-e486a4bace3c/volumes" Oct 11 03:24:47 crc kubenswrapper[4754]: I1011 03:24:47.313984 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 03:24:47 crc kubenswrapper[4754]: W1011 03:24:47.317685 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f98545_924e_4087_8f8e_15c9f9b260cb.slice/crio-3eaecc57ab39fdbf5c7db7d23ac1f774931c92d2f63e2c69d84f3d3697fb6f3a WatchSource:0}: Error finding container 3eaecc57ab39fdbf5c7db7d23ac1f774931c92d2f63e2c69d84f3d3697fb6f3a: Status 404 returned error can't find the container with id 3eaecc57ab39fdbf5c7db7d23ac1f774931c92d2f63e2c69d84f3d3697fb6f3a Oct 11 03:24:47 crc kubenswrapper[4754]: I1011 03:24:47.476697 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6f98545-924e-4087-8f8e-15c9f9b260cb","Type":"ContainerStarted","Data":"3eaecc57ab39fdbf5c7db7d23ac1f774931c92d2f63e2c69d84f3d3697fb6f3a"} Oct 11 03:24:48 crc kubenswrapper[4754]: I1011 03:24:48.487020 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6f98545-924e-4087-8f8e-15c9f9b260cb","Type":"ContainerStarted","Data":"0de8ba46cc5587b547f8a8ec908b2473fe0e37389adada435b7dec5f13046603"} Oct 11 03:24:48 crc kubenswrapper[4754]: I1011 03:24:48.516744 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.516725302 podStartE2EDuration="2.516725302s" podCreationTimestamp="2025-10-11 03:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:24:48.505455966 +0000 UTC m=+1136.064400761" watchObservedRunningTime="2025-10-11 03:24:48.516725302 +0000 UTC m=+1136.075670087" Oct 11 03:24:49 crc kubenswrapper[4754]: I1011 03:24:49.858655 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:24:49 crc kubenswrapper[4754]: I1011 03:24:49.858795 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 03:24:51 crc kubenswrapper[4754]: I1011 03:24:51.880381 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 03:24:54 crc kubenswrapper[4754]: I1011 03:24:54.859191 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 03:24:54 crc kubenswrapper[4754]: I1011 03:24:54.860316 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 03:24:54 crc kubenswrapper[4754]: I1011 03:24:54.865739 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:54 crc kubenswrapper[4754]: I1011 03:24:54.865773 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 03:24:55 crc kubenswrapper[4754]: I1011 03:24:55.878297 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cd491934-e354-499d-ad2d-f77049e8e4b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:55 crc kubenswrapper[4754]: I1011 03:24:55.878429 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cd491934-e354-499d-ad2d-f77049e8e4b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:55 crc kubenswrapper[4754]: I1011 03:24:55.903214 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:55 crc kubenswrapper[4754]: I1011 03:24:55.903551 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 03:24:56 crc kubenswrapper[4754]: I1011 03:24:56.772815 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 03:24:56 crc kubenswrapper[4754]: I1011 03:24:56.881419 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 11 03:24:56 crc kubenswrapper[4754]: I1011 03:24:56.917100 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 11 03:24:57 crc kubenswrapper[4754]: I1011 03:24:57.619133 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.867250 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.875691 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.875831 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.877352 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.877873 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.883144 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 03:25:04 crc kubenswrapper[4754]: I1011 03:25:04.888280 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 03:25:05 crc kubenswrapper[4754]: I1011 03:25:05.645876 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 03:25:05 crc kubenswrapper[4754]: I1011 03:25:05.651297 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 03:25:05 crc kubenswrapper[4754]: I1011 03:25:05.652182 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 03:25:13 crc kubenswrapper[4754]: I1011 03:25:13.873734 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:15 crc kubenswrapper[4754]: I1011 03:25:15.066934 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:18 crc kubenswrapper[4754]: I1011 03:25:18.031041 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="rabbitmq" containerID="cri-o://e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460" gracePeriod=604796 Oct 11 03:25:18 crc kubenswrapper[4754]: I1011 03:25:18.276823 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 11 03:25:19 crc kubenswrapper[4754]: I1011 03:25:19.393565 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="rabbitmq" containerID="cri-o://1961cf0c165b96b95951b30ff2ff71845df237418c6f67eeebab9cc5fb0291d3" gracePeriod=604796 Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.652247 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.800010 4754 generic.go:334] "Generic (PLEG): container finished" podID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerID="e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460" exitCode=0 Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.800055 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerDied","Data":"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460"} Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.800076 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.800093 4754 scope.go:117] "RemoveContainer" containerID="e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.800082 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a16af4b4-0dce-4af9-b87e-c9c80560d631","Type":"ContainerDied","Data":"30e8c792249feafbf5309f5a27f0723aec06e4585008c7a8b7d8b001788390f6"} Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.819506 4754 scope.go:117] "RemoveContainer" containerID="33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.847655 4754 scope.go:117] "RemoveContainer" containerID="e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460" Oct 11 03:25:24 crc kubenswrapper[4754]: E1011 03:25:24.848129 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460\": container with ID starting with e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460 not found: ID does not exist" containerID="e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.848167 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460"} err="failed to get container status \"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460\": rpc error: code = NotFound desc = could not find container \"e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460\": container with ID starting with e431ae187004abc37da8182ef10e49f12da0fc8cd5b4bfcf68d5b690584e3460 not found: ID does not exist" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.848186 4754 scope.go:117] "RemoveContainer" containerID="33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165" Oct 11 03:25:24 crc kubenswrapper[4754]: E1011 03:25:24.848617 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165\": container with ID starting with 33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165 not found: ID does not exist" containerID="33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.848638 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165"} err="failed to get container status \"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165\": rpc error: code = NotFound desc = could not find container \"33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165\": container with ID starting with 33e6dbd40f554adff4d201cbfee245fab6668860644d184343cb9e4089dca165 not found: ID does not exist" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851522 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851575 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d89xv\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851669 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851695 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851770 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851862 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851897 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851925 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.851953 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.852023 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.852055 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd\") pod \"a16af4b4-0dce-4af9-b87e-c9c80560d631\" (UID: \"a16af4b4-0dce-4af9-b87e-c9c80560d631\") " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.856936 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.857071 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.857202 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.860558 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info" (OuterVolumeSpecName: "pod-info") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.861231 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv" (OuterVolumeSpecName: "kube-api-access-d89xv") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "kube-api-access-d89xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.864126 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.868650 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.871174 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.884568 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data" (OuterVolumeSpecName: "config-data") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.933613 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf" (OuterVolumeSpecName: "server-conf") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956230 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956260 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-server-conf\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956287 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956331 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956343 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956353 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a16af4b4-0dce-4af9-b87e-c9c80560d631-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956361 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d89xv\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-kube-api-access-d89xv\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956369 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a16af4b4-0dce-4af9-b87e-c9c80560d631-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956377 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a16af4b4-0dce-4af9-b87e-c9c80560d631-pod-info\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.956403 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.967760 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a16af4b4-0dce-4af9-b87e-c9c80560d631" (UID: "a16af4b4-0dce-4af9-b87e-c9c80560d631"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:24 crc kubenswrapper[4754]: I1011 03:25:24.981849 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.058172 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.058216 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a16af4b4-0dce-4af9-b87e-c9c80560d631-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.135108 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.142349 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.157711 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:25 crc kubenswrapper[4754]: E1011 03:25:25.158229 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="rabbitmq" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.158246 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="rabbitmq" Oct 11 03:25:25 crc kubenswrapper[4754]: E1011 03:25:25.158278 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="setup-container" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.158284 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="setup-container" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.158439 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" containerName="rabbitmq" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.159331 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.161896 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162192 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162344 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162078 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9nkh9" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162130 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162512 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.162990 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.176552 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261275 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aff682f3-e7eb-4034-8e82-27a6dd09490a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261340 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261591 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261629 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261672 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aff682f3-e7eb-4034-8e82-27a6dd09490a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261863 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-config-data\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.261932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.262011 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.262041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.262070 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9lgm\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-kube-api-access-r9lgm\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364361 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-config-data\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364417 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364457 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364482 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364502 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9lgm\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-kube-api-access-r9lgm\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364554 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aff682f3-e7eb-4034-8e82-27a6dd09490a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364593 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364657 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.364729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aff682f3-e7eb-4034-8e82-27a6dd09490a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.365018 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.365524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.365944 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.366300 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-config-data\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.367222 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.368634 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.369040 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aff682f3-e7eb-4034-8e82-27a6dd09490a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.369496 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aff682f3-e7eb-4034-8e82-27a6dd09490a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.375074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aff682f3-e7eb-4034-8e82-27a6dd09490a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.375102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.393145 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9lgm\" (UniqueName: \"kubernetes.io/projected/aff682f3-e7eb-4034-8e82-27a6dd09490a-kube-api-access-r9lgm\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.413329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"aff682f3-e7eb-4034-8e82-27a6dd09490a\") " pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.491796 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.823212 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerID="1961cf0c165b96b95951b30ff2ff71845df237418c6f67eeebab9cc5fb0291d3" exitCode=0 Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.823494 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerDied","Data":"1961cf0c165b96b95951b30ff2ff71845df237418c6f67eeebab9cc5fb0291d3"} Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.940252 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:25 crc kubenswrapper[4754]: I1011 03:25:25.976851 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079646 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079755 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079779 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079825 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.079853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b98z8\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080117 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080166 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080214 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080235 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf\") pod \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\" (UID: \"ea4e5c84-3c0a-40fb-bccd-529f48f8c908\") " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.080950 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.082121 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.083073 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.084682 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8" (OuterVolumeSpecName: "kube-api-access-b98z8") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "kube-api-access-b98z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.087873 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.088434 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info" (OuterVolumeSpecName: "pod-info") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.088672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.090861 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.118785 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data" (OuterVolumeSpecName: "config-data") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.139536 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf" (OuterVolumeSpecName: "server-conf") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183726 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183763 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-pod-info\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183775 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183785 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183794 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-server-conf\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183801 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183809 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183817 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183824 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.183834 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b98z8\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-kube-api-access-b98z8\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.189657 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ea4e5c84-3c0a-40fb-bccd-529f48f8c908" (UID: "ea4e5c84-3c0a-40fb-bccd-529f48f8c908"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.205809 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.285179 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.285212 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea4e5c84-3c0a-40fb-bccd-529f48f8c908-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.838533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aff682f3-e7eb-4034-8e82-27a6dd09490a","Type":"ContainerStarted","Data":"ff08f0e15944db1b8b5ec526c8e525e1c3283ffa8d23529cb5c08606c69c1ddc"} Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.841487 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea4e5c84-3c0a-40fb-bccd-529f48f8c908","Type":"ContainerDied","Data":"4f2ca661286ccbe569ca591e36556b1d6e6ac265cce5b6381c6146f205575346"} Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.841560 4754 scope.go:117] "RemoveContainer" containerID="1961cf0c165b96b95951b30ff2ff71845df237418c6f67eeebab9cc5fb0291d3" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.841591 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.862427 4754 scope.go:117] "RemoveContainer" containerID="ae2d41dfd9071ca61bc08bf34b22249fad90e75f35f34c5be12366fc0525a845" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.884008 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.892711 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.919103 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:26 crc kubenswrapper[4754]: E1011 03:25:26.919627 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="rabbitmq" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.919657 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="rabbitmq" Oct 11 03:25:26 crc kubenswrapper[4754]: E1011 03:25:26.919687 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="setup-container" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.919698 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="setup-container" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.919945 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" containerName="rabbitmq" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.921813 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.925621 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.925680 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.925873 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.926091 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.926211 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.926318 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q5psc" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.926419 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 11 03:25:26 crc kubenswrapper[4754]: I1011 03:25:26.954102 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.094154 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a16af4b4-0dce-4af9-b87e-c9c80560d631" path="/var/lib/kubelet/pods/a16af4b4-0dce-4af9-b87e-c9c80560d631/volumes" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.095240 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4e5c84-3c0a-40fb-bccd-529f48f8c908" path="/var/lib/kubelet/pods/ea4e5c84-3c0a-40fb-bccd-529f48f8c908/volumes" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099043 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlrbt\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-kube-api-access-vlrbt\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099068 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099104 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099125 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099145 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099165 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099207 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.099431 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.200545 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlrbt\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-kube-api-access-vlrbt\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.200786 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.200858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.200981 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201065 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201151 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201227 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201445 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201643 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.201794 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.202068 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.202293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.202357 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.202711 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.203044 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.206274 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.206455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.211292 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.212053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.225237 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlrbt\" (UniqueName: \"kubernetes.io/projected/3b45e69c-6b94-4c29-8dcf-7a10c16a51be-kube-api-access-vlrbt\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.235526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3b45e69c-6b94-4c29-8dcf-7a10c16a51be\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.254444 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.840758 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 03:25:27 crc kubenswrapper[4754]: W1011 03:25:27.848710 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b45e69c_6b94_4c29_8dcf_7a10c16a51be.slice/crio-bb84be61fdb6b7f9a2f17d0ce09fe92816a255cd3a5eb80914997997e02091b6 WatchSource:0}: Error finding container bb84be61fdb6b7f9a2f17d0ce09fe92816a255cd3a5eb80914997997e02091b6: Status 404 returned error can't find the container with id bb84be61fdb6b7f9a2f17d0ce09fe92816a255cd3a5eb80914997997e02091b6 Oct 11 03:25:27 crc kubenswrapper[4754]: I1011 03:25:27.851142 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aff682f3-e7eb-4034-8e82-27a6dd09490a","Type":"ContainerStarted","Data":"c518316a680c0677bd7b7eefa6c3f0a1c7e2913814252c0606f5670c748379e7"} Oct 11 03:25:28 crc kubenswrapper[4754]: I1011 03:25:28.864560 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3b45e69c-6b94-4c29-8dcf-7a10c16a51be","Type":"ContainerStarted","Data":"bb84be61fdb6b7f9a2f17d0ce09fe92816a255cd3a5eb80914997997e02091b6"} Oct 11 03:25:28 crc kubenswrapper[4754]: I1011 03:25:28.929566 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:28 crc kubenswrapper[4754]: I1011 03:25:28.931589 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:28 crc kubenswrapper[4754]: I1011 03:25:28.934283 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 11 03:25:28 crc kubenswrapper[4754]: I1011 03:25:28.967677 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035028 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035139 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035160 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035174 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ck5h\" (UniqueName: \"kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035210 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.035247 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.136781 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.136858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.136915 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ck5h\" (UniqueName: \"kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.136943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.137509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.137590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.137836 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.138105 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.138152 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.138170 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.138744 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.171792 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ck5h\" (UniqueName: \"kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h\") pod \"dnsmasq-dns-6447ccbd8f-28c6g\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.267543 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.712242 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:29 crc kubenswrapper[4754]: W1011 03:25:29.712870 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36528aa7_f5f3_4055_864e_0473e7af1628.slice/crio-5ab97a256ae0ab90aa64b5b06c22236c45b17a48b540c1c56e5e3207160422f5 WatchSource:0}: Error finding container 5ab97a256ae0ab90aa64b5b06c22236c45b17a48b540c1c56e5e3207160422f5: Status 404 returned error can't find the container with id 5ab97a256ae0ab90aa64b5b06c22236c45b17a48b540c1c56e5e3207160422f5 Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.874574 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3b45e69c-6b94-4c29-8dcf-7a10c16a51be","Type":"ContainerStarted","Data":"a5be8d6a9d7e8223bcf0a4b2909224f52e8fb932c582aaa785fbbd105de0635d"} Oct 11 03:25:29 crc kubenswrapper[4754]: I1011 03:25:29.875914 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" event={"ID":"36528aa7-f5f3-4055-864e-0473e7af1628","Type":"ContainerStarted","Data":"5ab97a256ae0ab90aa64b5b06c22236c45b17a48b540c1c56e5e3207160422f5"} Oct 11 03:25:30 crc kubenswrapper[4754]: I1011 03:25:30.736094 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:25:30 crc kubenswrapper[4754]: I1011 03:25:30.736581 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:25:30 crc kubenswrapper[4754]: I1011 03:25:30.885067 4754 generic.go:334] "Generic (PLEG): container finished" podID="36528aa7-f5f3-4055-864e-0473e7af1628" containerID="97ffac4566fc8db3183bc436ba91ebe4db37de4e72bcb9022e70e250e807e401" exitCode=0 Oct 11 03:25:30 crc kubenswrapper[4754]: I1011 03:25:30.885188 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" event={"ID":"36528aa7-f5f3-4055-864e-0473e7af1628","Type":"ContainerDied","Data":"97ffac4566fc8db3183bc436ba91ebe4db37de4e72bcb9022e70e250e807e401"} Oct 11 03:25:31 crc kubenswrapper[4754]: I1011 03:25:31.907519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" event={"ID":"36528aa7-f5f3-4055-864e-0473e7af1628","Type":"ContainerStarted","Data":"6004980e5928f40d0cf4f3d4e067e42d8c3381cd44621826b0916aa76d577fc8"} Oct 11 03:25:31 crc kubenswrapper[4754]: I1011 03:25:31.907798 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:31 crc kubenswrapper[4754]: I1011 03:25:31.943248 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" podStartSLOduration=3.943205056 podStartE2EDuration="3.943205056s" podCreationTimestamp="2025-10-11 03:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:25:31.929262945 +0000 UTC m=+1179.488207790" watchObservedRunningTime="2025-10-11 03:25:31.943205056 +0000 UTC m=+1179.502149881" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.269156 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.334351 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.334630 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="dnsmasq-dns" containerID="cri-o://011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f" gracePeriod=10 Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.555670 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.557911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.581426 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641130 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641203 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb2tw\" (UniqueName: \"kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641278 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.641398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743159 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743238 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743281 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb2tw\" (UniqueName: \"kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743344 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743365 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.743384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.744227 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.744628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.744699 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.744749 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.745169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.761495 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb2tw\" (UniqueName: \"kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw\") pod \"dnsmasq-dns-864d5fc68c-879wl\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.804640 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.888907 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.946672 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config\") pod \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.946806 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc\") pod \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.946839 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb\") pod \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.946865 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb\") pod \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.946913 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr7sl\" (UniqueName: \"kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl\") pod \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\" (UID: \"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f\") " Oct 11 03:25:39 crc kubenswrapper[4754]: I1011 03:25:39.950682 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl" (OuterVolumeSpecName: "kube-api-access-sr7sl") pod "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" (UID: "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f"). InnerVolumeSpecName "kube-api-access-sr7sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:39.994277 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" (UID: "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.006475 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" (UID: "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.035277 4754 generic.go:334] "Generic (PLEG): container finished" podID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerID="011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f" exitCode=0 Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.035321 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" event={"ID":"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f","Type":"ContainerDied","Data":"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f"} Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.035355 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" event={"ID":"439b5343-acd0-4f95-a35c-ffbb3c6f7f7f","Type":"ContainerDied","Data":"fb8d96504028d4136d4e980c8a02fcd8be4f2cfc5de0edf19833ed50d2d0baa4"} Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.035370 4754 scope.go:117] "RemoveContainer" containerID="011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.035524 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-4lnrj" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.041502 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config" (OuterVolumeSpecName: "config") pod "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" (UID: "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.050522 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.050553 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.050569 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr7sl\" (UniqueName: \"kubernetes.io/projected/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-kube-api-access-sr7sl\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.050582 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.104122 4754 scope.go:117] "RemoveContainer" containerID="da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.114472 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" (UID: "439b5343-acd0-4f95-a35c-ffbb3c6f7f7f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.152744 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.198125 4754 scope.go:117] "RemoveContainer" containerID="011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f" Oct 11 03:25:40 crc kubenswrapper[4754]: E1011 03:25:40.198985 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f\": container with ID starting with 011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f not found: ID does not exist" containerID="011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.199016 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f"} err="failed to get container status \"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f\": rpc error: code = NotFound desc = could not find container \"011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f\": container with ID starting with 011b764bfdc4a94ca36d59c2255d7086844f0282dcbf637c6bd471b52a90724f not found: ID does not exist" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.199037 4754 scope.go:117] "RemoveContainer" containerID="da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d" Oct 11 03:25:40 crc kubenswrapper[4754]: E1011 03:25:40.199844 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d\": container with ID starting with da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d not found: ID does not exist" containerID="da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.199909 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d"} err="failed to get container status \"da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d\": rpc error: code = NotFound desc = could not find container \"da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d\": container with ID starting with da00935852dc6db1b4afabb695acafea87d781ee7018a619a58af4b5b58a1f5d not found: ID does not exist" Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.366888 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.375238 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-4lnrj"] Oct 11 03:25:40 crc kubenswrapper[4754]: I1011 03:25:40.440770 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:25:40 crc kubenswrapper[4754]: W1011 03:25:40.448650 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6098849_4d4b_4519_8e0e_548914f42c85.slice/crio-0701c77c05bdecf3682c2b3cc2647c4c02fca10901d173b98f2d1d54cda31224 WatchSource:0}: Error finding container 0701c77c05bdecf3682c2b3cc2647c4c02fca10901d173b98f2d1d54cda31224: Status 404 returned error can't find the container with id 0701c77c05bdecf3682c2b3cc2647c4c02fca10901d173b98f2d1d54cda31224 Oct 11 03:25:41 crc kubenswrapper[4754]: I1011 03:25:41.047741 4754 generic.go:334] "Generic (PLEG): container finished" podID="e6098849-4d4b-4519-8e0e-548914f42c85" containerID="41995a927cc6e9d1f0eaefab46648296e3bd41164adf1f9970e6b4c99b13c406" exitCode=0 Oct 11 03:25:41 crc kubenswrapper[4754]: I1011 03:25:41.047794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" event={"ID":"e6098849-4d4b-4519-8e0e-548914f42c85","Type":"ContainerDied","Data":"41995a927cc6e9d1f0eaefab46648296e3bd41164adf1f9970e6b4c99b13c406"} Oct 11 03:25:41 crc kubenswrapper[4754]: I1011 03:25:41.048276 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" event={"ID":"e6098849-4d4b-4519-8e0e-548914f42c85","Type":"ContainerStarted","Data":"0701c77c05bdecf3682c2b3cc2647c4c02fca10901d173b98f2d1d54cda31224"} Oct 11 03:25:41 crc kubenswrapper[4754]: I1011 03:25:41.101100 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" path="/var/lib/kubelet/pods/439b5343-acd0-4f95-a35c-ffbb3c6f7f7f/volumes" Oct 11 03:25:42 crc kubenswrapper[4754]: I1011 03:25:42.060719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" event={"ID":"e6098849-4d4b-4519-8e0e-548914f42c85","Type":"ContainerStarted","Data":"6f1d86e44f592c3e7d660121f5138a79c37b9c4478384c5e4e2532841acf6310"} Oct 11 03:25:42 crc kubenswrapper[4754]: I1011 03:25:42.061059 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:42 crc kubenswrapper[4754]: I1011 03:25:42.080672 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" podStartSLOduration=3.080653846 podStartE2EDuration="3.080653846s" podCreationTimestamp="2025-10-11 03:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:25:42.076248953 +0000 UTC m=+1189.635193758" watchObservedRunningTime="2025-10-11 03:25:42.080653846 +0000 UTC m=+1189.639598631" Oct 11 03:25:49 crc kubenswrapper[4754]: I1011 03:25:49.891182 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:25:49 crc kubenswrapper[4754]: I1011 03:25:49.992703 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:49 crc kubenswrapper[4754]: I1011 03:25:49.993079 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="dnsmasq-dns" containerID="cri-o://6004980e5928f40d0cf4f3d4e067e42d8c3381cd44621826b0916aa76d577fc8" gracePeriod=10 Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.157487 4754 generic.go:334] "Generic (PLEG): container finished" podID="36528aa7-f5f3-4055-864e-0473e7af1628" containerID="6004980e5928f40d0cf4f3d4e067e42d8c3381cd44621826b0916aa76d577fc8" exitCode=0 Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.157615 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" event={"ID":"36528aa7-f5f3-4055-864e-0473e7af1628","Type":"ContainerDied","Data":"6004980e5928f40d0cf4f3d4e067e42d8c3381cd44621826b0916aa76d577fc8"} Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.455033 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562412 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562481 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562585 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562675 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ck5h\" (UniqueName: \"kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562735 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.562752 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb\") pod \"36528aa7-f5f3-4055-864e-0473e7af1628\" (UID: \"36528aa7-f5f3-4055-864e-0473e7af1628\") " Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.568136 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h" (OuterVolumeSpecName: "kube-api-access-9ck5h") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "kube-api-access-9ck5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.615413 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.616459 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.616484 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.618062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.620247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config" (OuterVolumeSpecName: "config") pod "36528aa7-f5f3-4055-864e-0473e7af1628" (UID: "36528aa7-f5f3-4055-864e-0473e7af1628"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665260 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ck5h\" (UniqueName: \"kubernetes.io/projected/36528aa7-f5f3-4055-864e-0473e7af1628-kube-api-access-9ck5h\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665294 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665303 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665315 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665324 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:50 crc kubenswrapper[4754]: I1011 03:25:50.665332 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36528aa7-f5f3-4055-864e-0473e7af1628-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.167939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" event={"ID":"36528aa7-f5f3-4055-864e-0473e7af1628","Type":"ContainerDied","Data":"5ab97a256ae0ab90aa64b5b06c22236c45b17a48b540c1c56e5e3207160422f5"} Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.169005 4754 scope.go:117] "RemoveContainer" containerID="6004980e5928f40d0cf4f3d4e067e42d8c3381cd44621826b0916aa76d577fc8" Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.168037 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-28c6g" Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.196380 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.200220 4754 scope.go:117] "RemoveContainer" containerID="97ffac4566fc8db3183bc436ba91ebe4db37de4e72bcb9022e70e250e807e401" Oct 11 03:25:51 crc kubenswrapper[4754]: I1011 03:25:51.203809 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-28c6g"] Oct 11 03:25:53 crc kubenswrapper[4754]: I1011 03:25:53.098076 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" path="/var/lib/kubelet/pods/36528aa7-f5f3-4055-864e-0473e7af1628/volumes" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.100823 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6"] Oct 11 03:26:00 crc kubenswrapper[4754]: E1011 03:26:00.101832 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.101847 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: E1011 03:26:00.101867 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="init" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.101876 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="init" Oct 11 03:26:00 crc kubenswrapper[4754]: E1011 03:26:00.101894 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.101905 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: E1011 03:26:00.101949 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="init" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.101957 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="init" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.104158 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="36528aa7-f5f3-4055-864e-0473e7af1628" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.104339 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="439b5343-acd0-4f95-a35c-ffbb3c6f7f7f" containerName="dnsmasq-dns" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.105114 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.112113 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.112483 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.112739 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.113527 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.135893 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6"] Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.243828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.243967 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.244059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84rn\" (UniqueName: \"kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.244129 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.247597 4754 generic.go:334] "Generic (PLEG): container finished" podID="aff682f3-e7eb-4034-8e82-27a6dd09490a" containerID="c518316a680c0677bd7b7eefa6c3f0a1c7e2913814252c0606f5670c748379e7" exitCode=0 Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.247647 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aff682f3-e7eb-4034-8e82-27a6dd09490a","Type":"ContainerDied","Data":"c518316a680c0677bd7b7eefa6c3f0a1c7e2913814252c0606f5670c748379e7"} Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.345285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84rn\" (UniqueName: \"kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.345400 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.345495 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.345638 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.350374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.352904 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.352951 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.367056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84rn\" (UniqueName: \"kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.435935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.736655 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.737017 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.992077 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6"] Oct 11 03:26:00 crc kubenswrapper[4754]: I1011 03:26:00.996460 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:26:01 crc kubenswrapper[4754]: I1011 03:26:01.257086 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" event={"ID":"f009017c-d229-4243-922f-cd0ded520f61","Type":"ContainerStarted","Data":"28689a853d3fa75c11cf66143b83732514a3ebdc2f66ad98da8e9e2ffeb81f81"} Oct 11 03:26:01 crc kubenswrapper[4754]: I1011 03:26:01.259391 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aff682f3-e7eb-4034-8e82-27a6dd09490a","Type":"ContainerStarted","Data":"de6e55c9c147198408524ae85a5f5b8d37505854eba7df4142df0b4656293b11"} Oct 11 03:26:01 crc kubenswrapper[4754]: I1011 03:26:01.259547 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 11 03:26:01 crc kubenswrapper[4754]: I1011 03:26:01.290805 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.290777338 podStartE2EDuration="36.290777338s" podCreationTimestamp="2025-10-11 03:25:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:26:01.284559374 +0000 UTC m=+1208.843504179" watchObservedRunningTime="2025-10-11 03:26:01.290777338 +0000 UTC m=+1208.849722123" Oct 11 03:26:02 crc kubenswrapper[4754]: I1011 03:26:02.268687 4754 generic.go:334] "Generic (PLEG): container finished" podID="3b45e69c-6b94-4c29-8dcf-7a10c16a51be" containerID="a5be8d6a9d7e8223bcf0a4b2909224f52e8fb932c582aaa785fbbd105de0635d" exitCode=0 Oct 11 03:26:02 crc kubenswrapper[4754]: I1011 03:26:02.268879 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3b45e69c-6b94-4c29-8dcf-7a10c16a51be","Type":"ContainerDied","Data":"a5be8d6a9d7e8223bcf0a4b2909224f52e8fb932c582aaa785fbbd105de0635d"} Oct 11 03:26:03 crc kubenswrapper[4754]: I1011 03:26:03.281525 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3b45e69c-6b94-4c29-8dcf-7a10c16a51be","Type":"ContainerStarted","Data":"27e30b40a941f5ef32b755ef774d54613936f16395b303ab47864895d90ba5bd"} Oct 11 03:26:03 crc kubenswrapper[4754]: I1011 03:26:03.282417 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:26:03 crc kubenswrapper[4754]: I1011 03:26:03.321933 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.321913393 podStartE2EDuration="37.321913393s" podCreationTimestamp="2025-10-11 03:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:26:03.312823548 +0000 UTC m=+1210.871768333" watchObservedRunningTime="2025-10-11 03:26:03.321913393 +0000 UTC m=+1210.880858178" Oct 11 03:26:10 crc kubenswrapper[4754]: I1011 03:26:10.339921 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" event={"ID":"f009017c-d229-4243-922f-cd0ded520f61","Type":"ContainerStarted","Data":"760df319ae572dd7e0b2d876932b44166a3b0a1e1a5b71f252c89e576bf75357"} Oct 11 03:26:15 crc kubenswrapper[4754]: I1011 03:26:15.494107 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 11 03:26:15 crc kubenswrapper[4754]: I1011 03:26:15.528507 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" podStartSLOduration=6.99930651 podStartE2EDuration="15.528483672s" podCreationTimestamp="2025-10-11 03:26:00 +0000 UTC" firstStartedPulling="2025-10-11 03:26:00.996247261 +0000 UTC m=+1208.555192046" lastFinishedPulling="2025-10-11 03:26:09.525424423 +0000 UTC m=+1217.084369208" observedRunningTime="2025-10-11 03:26:10.360520615 +0000 UTC m=+1217.919465400" watchObservedRunningTime="2025-10-11 03:26:15.528483672 +0000 UTC m=+1223.087428457" Oct 11 03:26:17 crc kubenswrapper[4754]: I1011 03:26:17.259559 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 11 03:26:21 crc kubenswrapper[4754]: I1011 03:26:21.426771 4754 generic.go:334] "Generic (PLEG): container finished" podID="f009017c-d229-4243-922f-cd0ded520f61" containerID="760df319ae572dd7e0b2d876932b44166a3b0a1e1a5b71f252c89e576bf75357" exitCode=0 Oct 11 03:26:21 crc kubenswrapper[4754]: I1011 03:26:21.426898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" event={"ID":"f009017c-d229-4243-922f-cd0ded520f61","Type":"ContainerDied","Data":"760df319ae572dd7e0b2d876932b44166a3b0a1e1a5b71f252c89e576bf75357"} Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.814085 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.933497 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s84rn\" (UniqueName: \"kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn\") pod \"f009017c-d229-4243-922f-cd0ded520f61\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.933740 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory\") pod \"f009017c-d229-4243-922f-cd0ded520f61\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.933777 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key\") pod \"f009017c-d229-4243-922f-cd0ded520f61\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.933818 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle\") pod \"f009017c-d229-4243-922f-cd0ded520f61\" (UID: \"f009017c-d229-4243-922f-cd0ded520f61\") " Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.939297 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f009017c-d229-4243-922f-cd0ded520f61" (UID: "f009017c-d229-4243-922f-cd0ded520f61"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.939714 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn" (OuterVolumeSpecName: "kube-api-access-s84rn") pod "f009017c-d229-4243-922f-cd0ded520f61" (UID: "f009017c-d229-4243-922f-cd0ded520f61"). InnerVolumeSpecName "kube-api-access-s84rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.959249 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f009017c-d229-4243-922f-cd0ded520f61" (UID: "f009017c-d229-4243-922f-cd0ded520f61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:26:22 crc kubenswrapper[4754]: I1011 03:26:22.960866 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory" (OuterVolumeSpecName: "inventory") pod "f009017c-d229-4243-922f-cd0ded520f61" (UID: "f009017c-d229-4243-922f-cd0ded520f61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.036232 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s84rn\" (UniqueName: \"kubernetes.io/projected/f009017c-d229-4243-922f-cd0ded520f61-kube-api-access-s84rn\") on node \"crc\" DevicePath \"\"" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.036260 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.036268 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.036277 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f009017c-d229-4243-922f-cd0ded520f61-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.444711 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" event={"ID":"f009017c-d229-4243-922f-cd0ded520f61","Type":"ContainerDied","Data":"28689a853d3fa75c11cf66143b83732514a3ebdc2f66ad98da8e9e2ffeb81f81"} Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.444751 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28689a853d3fa75c11cf66143b83732514a3ebdc2f66ad98da8e9e2ffeb81f81" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.444776 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.536182 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk"] Oct 11 03:26:23 crc kubenswrapper[4754]: E1011 03:26:23.536616 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f009017c-d229-4243-922f-cd0ded520f61" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.536640 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f009017c-d229-4243-922f-cd0ded520f61" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.536899 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f009017c-d229-4243-922f-cd0ded520f61" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.537631 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.541013 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.541198 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.541229 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.546639 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.546701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.546795 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.546874 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsnqv\" (UniqueName: \"kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.547579 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.551307 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk"] Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.647995 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.648103 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsnqv\" (UniqueName: \"kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.648204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.648238 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.652749 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.652831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.656620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.664748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsnqv\" (UniqueName: \"kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:23 crc kubenswrapper[4754]: I1011 03:26:23.860541 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:26:24 crc kubenswrapper[4754]: I1011 03:26:24.351813 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk"] Oct 11 03:26:24 crc kubenswrapper[4754]: I1011 03:26:24.453584 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" event={"ID":"53d3017f-d07c-4e07-b696-0ce8a6bc8e71","Type":"ContainerStarted","Data":"ab69c575af754976f4c7580bf65ffb38f9c51177ef7989cf3d553a6e58cf6e97"} Oct 11 03:26:25 crc kubenswrapper[4754]: I1011 03:26:25.482744 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" event={"ID":"53d3017f-d07c-4e07-b696-0ce8a6bc8e71","Type":"ContainerStarted","Data":"f66067f6561342ebf8ff8fffea0fc6e2dc47748a8131518c4db4ddb1f6e241c6"} Oct 11 03:26:25 crc kubenswrapper[4754]: I1011 03:26:25.503730 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" podStartSLOduration=2.077770564 podStartE2EDuration="2.503713662s" podCreationTimestamp="2025-10-11 03:26:23 +0000 UTC" firstStartedPulling="2025-10-11 03:26:24.358059377 +0000 UTC m=+1231.917004172" lastFinishedPulling="2025-10-11 03:26:24.784002485 +0000 UTC m=+1232.342947270" observedRunningTime="2025-10-11 03:26:25.502332153 +0000 UTC m=+1233.061276938" watchObservedRunningTime="2025-10-11 03:26:25.503713662 +0000 UTC m=+1233.062658437" Oct 11 03:26:30 crc kubenswrapper[4754]: I1011 03:26:30.736750 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:26:30 crc kubenswrapper[4754]: I1011 03:26:30.737737 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:26:30 crc kubenswrapper[4754]: I1011 03:26:30.737812 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:26:30 crc kubenswrapper[4754]: I1011 03:26:30.740730 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:26:30 crc kubenswrapper[4754]: I1011 03:26:30.740933 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9" gracePeriod=600 Oct 11 03:26:31 crc kubenswrapper[4754]: I1011 03:26:31.539761 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9" exitCode=0 Oct 11 03:26:31 crc kubenswrapper[4754]: I1011 03:26:31.539785 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9"} Oct 11 03:26:31 crc kubenswrapper[4754]: I1011 03:26:31.541293 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6"} Oct 11 03:26:31 crc kubenswrapper[4754]: I1011 03:26:31.541384 4754 scope.go:117] "RemoveContainer" containerID="f051113b40ef9d35aae42e161ed38132580580575011913971578373c9654494" Oct 11 03:28:03 crc kubenswrapper[4754]: I1011 03:28:03.364649 4754 scope.go:117] "RemoveContainer" containerID="6ea1330af2899f86cb736412876e00f28b7fdee299340689c12a4dfe18c790ba" Oct 11 03:28:03 crc kubenswrapper[4754]: I1011 03:28:03.425385 4754 scope.go:117] "RemoveContainer" containerID="075b4884ebe3d92fcdbaac365a00f85b096e3b6f5d832b1b7bbb3613dc8f4942" Oct 11 03:28:03 crc kubenswrapper[4754]: I1011 03:28:03.457689 4754 scope.go:117] "RemoveContainer" containerID="30be761ee00c4a2c43fd00e2b0c4876fccc0fe16a7790f354cbc284fc09e4f55" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.703533 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.706698 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.717200 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.733823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.734161 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.734550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qshpf\" (UniqueName: \"kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.835705 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.835939 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qshpf\" (UniqueName: \"kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.836042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.836732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.836938 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:44 crc kubenswrapper[4754]: I1011 03:28:44.859822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qshpf\" (UniqueName: \"kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf\") pod \"certified-operators-pjlj4\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:45 crc kubenswrapper[4754]: I1011 03:28:45.037819 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:45 crc kubenswrapper[4754]: I1011 03:28:45.555355 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:45 crc kubenswrapper[4754]: I1011 03:28:45.781902 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerStarted","Data":"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707"} Oct 11 03:28:45 crc kubenswrapper[4754]: I1011 03:28:45.782245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerStarted","Data":"6a4eb21165c1f36a31af27723def6ee94e71422ccc462bec45e983b42e434d7e"} Oct 11 03:28:46 crc kubenswrapper[4754]: I1011 03:28:46.791036 4754 generic.go:334] "Generic (PLEG): container finished" podID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerID="fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707" exitCode=0 Oct 11 03:28:46 crc kubenswrapper[4754]: I1011 03:28:46.791135 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerDied","Data":"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707"} Oct 11 03:28:47 crc kubenswrapper[4754]: I1011 03:28:47.801629 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerStarted","Data":"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4"} Oct 11 03:28:48 crc kubenswrapper[4754]: I1011 03:28:48.816636 4754 generic.go:334] "Generic (PLEG): container finished" podID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerID="447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4" exitCode=0 Oct 11 03:28:48 crc kubenswrapper[4754]: I1011 03:28:48.816679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerDied","Data":"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4"} Oct 11 03:28:50 crc kubenswrapper[4754]: I1011 03:28:50.834579 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerStarted","Data":"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06"} Oct 11 03:28:50 crc kubenswrapper[4754]: I1011 03:28:50.855520 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pjlj4" podStartSLOduration=4.299652952 podStartE2EDuration="6.855501581s" podCreationTimestamp="2025-10-11 03:28:44 +0000 UTC" firstStartedPulling="2025-10-11 03:28:46.792883647 +0000 UTC m=+1374.351828432" lastFinishedPulling="2025-10-11 03:28:49.348732276 +0000 UTC m=+1376.907677061" observedRunningTime="2025-10-11 03:28:50.855378028 +0000 UTC m=+1378.414322813" watchObservedRunningTime="2025-10-11 03:28:50.855501581 +0000 UTC m=+1378.414446356" Oct 11 03:28:55 crc kubenswrapper[4754]: I1011 03:28:55.038590 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:55 crc kubenswrapper[4754]: I1011 03:28:55.039260 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:55 crc kubenswrapper[4754]: I1011 03:28:55.083576 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:55 crc kubenswrapper[4754]: I1011 03:28:55.919290 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:55 crc kubenswrapper[4754]: I1011 03:28:55.968855 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:57 crc kubenswrapper[4754]: I1011 03:28:57.894404 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pjlj4" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="registry-server" containerID="cri-o://ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06" gracePeriod=2 Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.341995 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.499615 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qshpf\" (UniqueName: \"kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf\") pod \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.499746 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content\") pod \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.499811 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities\") pod \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\" (UID: \"8b4fbfb3-dafc-479e-8729-1f1a1ff80153\") " Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.500998 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities" (OuterVolumeSpecName: "utilities") pod "8b4fbfb3-dafc-479e-8729-1f1a1ff80153" (UID: "8b4fbfb3-dafc-479e-8729-1f1a1ff80153"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.507162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf" (OuterVolumeSpecName: "kube-api-access-qshpf") pod "8b4fbfb3-dafc-479e-8729-1f1a1ff80153" (UID: "8b4fbfb3-dafc-479e-8729-1f1a1ff80153"). InnerVolumeSpecName "kube-api-access-qshpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.542128 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b4fbfb3-dafc-479e-8729-1f1a1ff80153" (UID: "8b4fbfb3-dafc-479e-8729-1f1a1ff80153"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.601553 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qshpf\" (UniqueName: \"kubernetes.io/projected/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-kube-api-access-qshpf\") on node \"crc\" DevicePath \"\"" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.601591 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.601600 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4fbfb3-dafc-479e-8729-1f1a1ff80153-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.906467 4754 generic.go:334] "Generic (PLEG): container finished" podID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerID="ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06" exitCode=0 Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.906528 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerDied","Data":"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06"} Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.906547 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjlj4" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.906694 4754 scope.go:117] "RemoveContainer" containerID="ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.906706 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjlj4" event={"ID":"8b4fbfb3-dafc-479e-8729-1f1a1ff80153","Type":"ContainerDied","Data":"6a4eb21165c1f36a31af27723def6ee94e71422ccc462bec45e983b42e434d7e"} Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.945554 4754 scope.go:117] "RemoveContainer" containerID="447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4" Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.959088 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.968646 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pjlj4"] Oct 11 03:28:58 crc kubenswrapper[4754]: I1011 03:28:58.974737 4754 scope.go:117] "RemoveContainer" containerID="fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.013370 4754 scope.go:117] "RemoveContainer" containerID="ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06" Oct 11 03:28:59 crc kubenswrapper[4754]: E1011 03:28:59.013906 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06\": container with ID starting with ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06 not found: ID does not exist" containerID="ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.013956 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06"} err="failed to get container status \"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06\": rpc error: code = NotFound desc = could not find container \"ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06\": container with ID starting with ddd718d36283dcb8c8a2231282a2de4578d5f54da4ef6bdf750fdc30090f1f06 not found: ID does not exist" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.014004 4754 scope.go:117] "RemoveContainer" containerID="447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4" Oct 11 03:28:59 crc kubenswrapper[4754]: E1011 03:28:59.014434 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4\": container with ID starting with 447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4 not found: ID does not exist" containerID="447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.014482 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4"} err="failed to get container status \"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4\": rpc error: code = NotFound desc = could not find container \"447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4\": container with ID starting with 447a93ec31c742708b0a7393fb2790ef99e718bf5d1ddc20390184672058b0d4 not found: ID does not exist" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.014527 4754 scope.go:117] "RemoveContainer" containerID="fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707" Oct 11 03:28:59 crc kubenswrapper[4754]: E1011 03:28:59.014783 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707\": container with ID starting with fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707 not found: ID does not exist" containerID="fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.014812 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707"} err="failed to get container status \"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707\": rpc error: code = NotFound desc = could not find container \"fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707\": container with ID starting with fca481cdffc69116e2498220c5474c413c8c41bff93616fa1cf898438c7f1707 not found: ID does not exist" Oct 11 03:28:59 crc kubenswrapper[4754]: I1011 03:28:59.095245 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" path="/var/lib/kubelet/pods/8b4fbfb3-dafc-479e-8729-1f1a1ff80153/volumes" Oct 11 03:28:59 crc kubenswrapper[4754]: E1011 03:28:59.151377 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b4fbfb3_dafc_479e_8729_1f1a1ff80153.slice/crio-6a4eb21165c1f36a31af27723def6ee94e71422ccc462bec45e983b42e434d7e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b4fbfb3_dafc_479e_8729_1f1a1ff80153.slice\": RecentStats: unable to find data in memory cache]" Oct 11 03:29:00 crc kubenswrapper[4754]: I1011 03:29:00.736061 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:29:00 crc kubenswrapper[4754]: I1011 03:29:00.736370 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:29:29 crc kubenswrapper[4754]: E1011 03:29:29.868315 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53d3017f_d07c_4e07_b696_0ce8a6bc8e71.slice/crio-f66067f6561342ebf8ff8fffea0fc6e2dc47748a8131518c4db4ddb1f6e241c6.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.094011 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:30 crc kubenswrapper[4754]: E1011 03:29:30.094457 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="extract-content" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.094487 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="extract-content" Oct 11 03:29:30 crc kubenswrapper[4754]: E1011 03:29:30.094529 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="extract-utilities" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.094539 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="extract-utilities" Oct 11 03:29:30 crc kubenswrapper[4754]: E1011 03:29:30.094557 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="registry-server" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.094563 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="registry-server" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.094753 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b4fbfb3-dafc-479e-8729-1f1a1ff80153" containerName="registry-server" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.096269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.106031 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.164019 4754 generic.go:334] "Generic (PLEG): container finished" podID="53d3017f-d07c-4e07-b696-0ce8a6bc8e71" containerID="f66067f6561342ebf8ff8fffea0fc6e2dc47748a8131518c4db4ddb1f6e241c6" exitCode=0 Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.164057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" event={"ID":"53d3017f-d07c-4e07-b696-0ce8a6bc8e71","Type":"ContainerDied","Data":"f66067f6561342ebf8ff8fffea0fc6e2dc47748a8131518c4db4ddb1f6e241c6"} Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.251554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.251604 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx5dt\" (UniqueName: \"kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.251631 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.353059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.353396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx5dt\" (UniqueName: \"kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.353514 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.353645 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.353939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.376981 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx5dt\" (UniqueName: \"kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt\") pod \"redhat-marketplace-t2mtz\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.426747 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.736310 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.736362 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:29:30 crc kubenswrapper[4754]: I1011 03:29:30.918221 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.175357 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8754957-fb20-421e-9d8d-33bc573a33af" containerID="db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1" exitCode=0 Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.175448 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerDied","Data":"db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1"} Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.177116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerStarted","Data":"f3c98db5cb7602e69f3bec7f5842a5132ed61d725185b73a930c1b535e564e02"} Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.556801 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.678584 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle\") pod \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.678651 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory\") pod \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.678692 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key\") pod \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.678804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsnqv\" (UniqueName: \"kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv\") pod \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\" (UID: \"53d3017f-d07c-4e07-b696-0ce8a6bc8e71\") " Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.684722 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "53d3017f-d07c-4e07-b696-0ce8a6bc8e71" (UID: "53d3017f-d07c-4e07-b696-0ce8a6bc8e71"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.685045 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv" (OuterVolumeSpecName: "kube-api-access-lsnqv") pod "53d3017f-d07c-4e07-b696-0ce8a6bc8e71" (UID: "53d3017f-d07c-4e07-b696-0ce8a6bc8e71"). InnerVolumeSpecName "kube-api-access-lsnqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.706329 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53d3017f-d07c-4e07-b696-0ce8a6bc8e71" (UID: "53d3017f-d07c-4e07-b696-0ce8a6bc8e71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.708676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory" (OuterVolumeSpecName: "inventory") pod "53d3017f-d07c-4e07-b696-0ce8a6bc8e71" (UID: "53d3017f-d07c-4e07-b696-0ce8a6bc8e71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.781666 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.781707 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.781720 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsnqv\" (UniqueName: \"kubernetes.io/projected/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-kube-api-access-lsnqv\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:31 crc kubenswrapper[4754]: I1011 03:29:31.781735 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d3017f-d07c-4e07-b696-0ce8a6bc8e71-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.185071 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8754957-fb20-421e-9d8d-33bc573a33af" containerID="2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44" exitCode=0 Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.185404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerDied","Data":"2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44"} Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.187297 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" event={"ID":"53d3017f-d07c-4e07-b696-0ce8a6bc8e71","Type":"ContainerDied","Data":"ab69c575af754976f4c7580bf65ffb38f9c51177ef7989cf3d553a6e58cf6e97"} Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.187325 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab69c575af754976f4c7580bf65ffb38f9c51177ef7989cf3d553a6e58cf6e97" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.187358 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.270742 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh"] Oct 11 03:29:32 crc kubenswrapper[4754]: E1011 03:29:32.271103 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d3017f-d07c-4e07-b696-0ce8a6bc8e71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.271116 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d3017f-d07c-4e07-b696-0ce8a6bc8e71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.271293 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d3017f-d07c-4e07-b696-0ce8a6bc8e71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.271881 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.274563 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.274712 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.274845 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.275436 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.285063 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh"] Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.394312 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.394367 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p28qd\" (UniqueName: \"kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.394461 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.496003 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.496096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.496132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p28qd\" (UniqueName: \"kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.501849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.502021 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.516287 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p28qd\" (UniqueName: \"kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-blnfh\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:32 crc kubenswrapper[4754]: I1011 03:29:32.595289 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:29:33 crc kubenswrapper[4754]: I1011 03:29:33.132774 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh"] Oct 11 03:29:33 crc kubenswrapper[4754]: W1011 03:29:33.136039 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2bd066_1dae_406a_8a53_28f620db3fdf.slice/crio-0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45 WatchSource:0}: Error finding container 0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45: Status 404 returned error can't find the container with id 0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45 Oct 11 03:29:33 crc kubenswrapper[4754]: I1011 03:29:33.195469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" event={"ID":"3f2bd066-1dae-406a-8a53-28f620db3fdf","Type":"ContainerStarted","Data":"0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45"} Oct 11 03:29:33 crc kubenswrapper[4754]: I1011 03:29:33.199007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerStarted","Data":"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210"} Oct 11 03:29:33 crc kubenswrapper[4754]: I1011 03:29:33.219649 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t2mtz" podStartSLOduration=1.693976108 podStartE2EDuration="3.219628199s" podCreationTimestamp="2025-10-11 03:29:30 +0000 UTC" firstStartedPulling="2025-10-11 03:29:31.17646606 +0000 UTC m=+1418.735410845" lastFinishedPulling="2025-10-11 03:29:32.702118151 +0000 UTC m=+1420.261062936" observedRunningTime="2025-10-11 03:29:33.217038155 +0000 UTC m=+1420.775982960" watchObservedRunningTime="2025-10-11 03:29:33.219628199 +0000 UTC m=+1420.778572984" Oct 11 03:29:34 crc kubenswrapper[4754]: I1011 03:29:34.214628 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" event={"ID":"3f2bd066-1dae-406a-8a53-28f620db3fdf","Type":"ContainerStarted","Data":"2aa328f4587bd88700843fbaa8054dd6d9e54a35447c252f6e51c1c2e3e47af3"} Oct 11 03:29:34 crc kubenswrapper[4754]: I1011 03:29:34.248554 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" podStartSLOduration=1.7527826 podStartE2EDuration="2.248518401s" podCreationTimestamp="2025-10-11 03:29:32 +0000 UTC" firstStartedPulling="2025-10-11 03:29:33.139015572 +0000 UTC m=+1420.697960357" lastFinishedPulling="2025-10-11 03:29:33.634751343 +0000 UTC m=+1421.193696158" observedRunningTime="2025-10-11 03:29:34.236023616 +0000 UTC m=+1421.794968401" watchObservedRunningTime="2025-10-11 03:29:34.248518401 +0000 UTC m=+1421.807463226" Oct 11 03:29:40 crc kubenswrapper[4754]: I1011 03:29:40.428311 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:40 crc kubenswrapper[4754]: I1011 03:29:40.428803 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:40 crc kubenswrapper[4754]: I1011 03:29:40.484391 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:41 crc kubenswrapper[4754]: I1011 03:29:41.320909 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:41 crc kubenswrapper[4754]: I1011 03:29:41.376612 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.293583 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t2mtz" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="registry-server" containerID="cri-o://1788589c2592061ef537631387349327e358012611fe4d223725694660b97210" gracePeriod=2 Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.718592 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.815853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx5dt\" (UniqueName: \"kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt\") pod \"c8754957-fb20-421e-9d8d-33bc573a33af\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.815922 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content\") pod \"c8754957-fb20-421e-9d8d-33bc573a33af\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.816145 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities\") pod \"c8754957-fb20-421e-9d8d-33bc573a33af\" (UID: \"c8754957-fb20-421e-9d8d-33bc573a33af\") " Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.816956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities" (OuterVolumeSpecName: "utilities") pod "c8754957-fb20-421e-9d8d-33bc573a33af" (UID: "c8754957-fb20-421e-9d8d-33bc573a33af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.821801 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt" (OuterVolumeSpecName: "kube-api-access-qx5dt") pod "c8754957-fb20-421e-9d8d-33bc573a33af" (UID: "c8754957-fb20-421e-9d8d-33bc573a33af"). InnerVolumeSpecName "kube-api-access-qx5dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.836589 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8754957-fb20-421e-9d8d-33bc573a33af" (UID: "c8754957-fb20-421e-9d8d-33bc573a33af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.918072 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.918101 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx5dt\" (UniqueName: \"kubernetes.io/projected/c8754957-fb20-421e-9d8d-33bc573a33af-kube-api-access-qx5dt\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:43 crc kubenswrapper[4754]: I1011 03:29:43.918113 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8754957-fb20-421e-9d8d-33bc573a33af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.309181 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8754957-fb20-421e-9d8d-33bc573a33af" containerID="1788589c2592061ef537631387349327e358012611fe4d223725694660b97210" exitCode=0 Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.309346 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerDied","Data":"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210"} Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.309801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2mtz" event={"ID":"c8754957-fb20-421e-9d8d-33bc573a33af","Type":"ContainerDied","Data":"f3c98db5cb7602e69f3bec7f5842a5132ed61d725185b73a930c1b535e564e02"} Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.309841 4754 scope.go:117] "RemoveContainer" containerID="1788589c2592061ef537631387349327e358012611fe4d223725694660b97210" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.309426 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2mtz" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.345139 4754 scope.go:117] "RemoveContainer" containerID="2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.363646 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.376429 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2mtz"] Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.391676 4754 scope.go:117] "RemoveContainer" containerID="db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.431585 4754 scope.go:117] "RemoveContainer" containerID="1788589c2592061ef537631387349327e358012611fe4d223725694660b97210" Oct 11 03:29:44 crc kubenswrapper[4754]: E1011 03:29:44.432081 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210\": container with ID starting with 1788589c2592061ef537631387349327e358012611fe4d223725694660b97210 not found: ID does not exist" containerID="1788589c2592061ef537631387349327e358012611fe4d223725694660b97210" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.432134 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210"} err="failed to get container status \"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210\": rpc error: code = NotFound desc = could not find container \"1788589c2592061ef537631387349327e358012611fe4d223725694660b97210\": container with ID starting with 1788589c2592061ef537631387349327e358012611fe4d223725694660b97210 not found: ID does not exist" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.432168 4754 scope.go:117] "RemoveContainer" containerID="2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44" Oct 11 03:29:44 crc kubenswrapper[4754]: E1011 03:29:44.432466 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44\": container with ID starting with 2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44 not found: ID does not exist" containerID="2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.432499 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44"} err="failed to get container status \"2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44\": rpc error: code = NotFound desc = could not find container \"2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44\": container with ID starting with 2e6f64ab2c8a223cf92c8a0756d6cb06ea297901196cbd089b80bd4b918f5f44 not found: ID does not exist" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.432519 4754 scope.go:117] "RemoveContainer" containerID="db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1" Oct 11 03:29:44 crc kubenswrapper[4754]: E1011 03:29:44.432725 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1\": container with ID starting with db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1 not found: ID does not exist" containerID="db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1" Oct 11 03:29:44 crc kubenswrapper[4754]: I1011 03:29:44.432744 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1"} err="failed to get container status \"db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1\": rpc error: code = NotFound desc = could not find container \"db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1\": container with ID starting with db2fb1a7354cdfa2b6e25bba483e4a9b4eff1c7a91b9fa22a3a7bd22ace8eaf1 not found: ID does not exist" Oct 11 03:29:45 crc kubenswrapper[4754]: I1011 03:29:45.092665 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" path="/var/lib/kubelet/pods/c8754957-fb20-421e-9d8d-33bc573a33af/volumes" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.791784 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:29:58 crc kubenswrapper[4754]: E1011 03:29:58.793231 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="extract-content" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.793257 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="extract-content" Oct 11 03:29:58 crc kubenswrapper[4754]: E1011 03:29:58.793286 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="extract-utilities" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.793299 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="extract-utilities" Oct 11 03:29:58 crc kubenswrapper[4754]: E1011 03:29:58.793331 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="registry-server" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.793343 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="registry-server" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.793702 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8754957-fb20-421e-9d8d-33bc573a33af" containerName="registry-server" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.796206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.830644 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.998313 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.998385 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:58 crc kubenswrapper[4754]: I1011 03:29:58.998704 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmd9n\" (UniqueName: \"kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.101454 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.100955 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.101554 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.102002 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.103849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmd9n\" (UniqueName: \"kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.135816 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmd9n\" (UniqueName: \"kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n\") pod \"community-operators-5h898\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.431252 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:29:59 crc kubenswrapper[4754]: I1011 03:29:59.858992 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.143778 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv"] Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.145070 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.147365 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.147699 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.151697 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv"] Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.322949 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgg4j\" (UniqueName: \"kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.323041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.323254 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.424942 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgg4j\" (UniqueName: \"kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.425017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.425083 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.426889 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.440591 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.447458 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgg4j\" (UniqueName: \"kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j\") pod \"collect-profiles-29335890-q8csv\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.464725 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.494027 4754 generic.go:334] "Generic (PLEG): container finished" podID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerID="563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69" exitCode=0 Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.494071 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerDied","Data":"563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69"} Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.494105 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerStarted","Data":"1042252d6e48e8941044bae7802e0da75e497bf29b74f371b793463f6aed1b6e"} Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.736540 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.736825 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.736869 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.737661 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.737722 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6" gracePeriod=600 Oct 11 03:30:00 crc kubenswrapper[4754]: I1011 03:30:00.990879 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv"] Oct 11 03:30:01 crc kubenswrapper[4754]: W1011 03:30:01.000401 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd842ba1d_77ef_4397_a138_849800aff4a9.slice/crio-81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00 WatchSource:0}: Error finding container 81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00: Status 404 returned error can't find the container with id 81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00 Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.503590 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6" exitCode=0 Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.503657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6"} Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.504145 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5"} Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.504165 4754 scope.go:117] "RemoveContainer" containerID="2e56d0dd96f801c2fc011b56201c9c11f3033cbdfae1ca75a5ef7775d1239af9" Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.508814 4754 generic.go:334] "Generic (PLEG): container finished" podID="d842ba1d-77ef-4397-a138-849800aff4a9" containerID="862993f30a12e79f8b52b2ac44dabc65baa9d1a7082832c6d3feb510cd1d0945" exitCode=0 Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.508862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" event={"ID":"d842ba1d-77ef-4397-a138-849800aff4a9","Type":"ContainerDied","Data":"862993f30a12e79f8b52b2ac44dabc65baa9d1a7082832c6d3feb510cd1d0945"} Oct 11 03:30:01 crc kubenswrapper[4754]: I1011 03:30:01.508892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" event={"ID":"d842ba1d-77ef-4397-a138-849800aff4a9","Type":"ContainerStarted","Data":"81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00"} Oct 11 03:30:02 crc kubenswrapper[4754]: I1011 03:30:02.519974 4754 generic.go:334] "Generic (PLEG): container finished" podID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerID="9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96" exitCode=0 Oct 11 03:30:02 crc kubenswrapper[4754]: I1011 03:30:02.520074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerDied","Data":"9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96"} Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:02.953573 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.011629 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgg4j\" (UniqueName: \"kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j\") pod \"d842ba1d-77ef-4397-a138-849800aff4a9\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.012076 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume\") pod \"d842ba1d-77ef-4397-a138-849800aff4a9\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.012280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume\") pod \"d842ba1d-77ef-4397-a138-849800aff4a9\" (UID: \"d842ba1d-77ef-4397-a138-849800aff4a9\") " Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.013257 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume" (OuterVolumeSpecName: "config-volume") pod "d842ba1d-77ef-4397-a138-849800aff4a9" (UID: "d842ba1d-77ef-4397-a138-849800aff4a9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.021179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d842ba1d-77ef-4397-a138-849800aff4a9" (UID: "d842ba1d-77ef-4397-a138-849800aff4a9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.021212 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j" (OuterVolumeSpecName: "kube-api-access-bgg4j") pod "d842ba1d-77ef-4397-a138-849800aff4a9" (UID: "d842ba1d-77ef-4397-a138-849800aff4a9"). InnerVolumeSpecName "kube-api-access-bgg4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.114468 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d842ba1d-77ef-4397-a138-849800aff4a9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.114488 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgg4j\" (UniqueName: \"kubernetes.io/projected/d842ba1d-77ef-4397-a138-849800aff4a9-kube-api-access-bgg4j\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.114498 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d842ba1d-77ef-4397-a138-849800aff4a9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.529157 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" event={"ID":"d842ba1d-77ef-4397-a138-849800aff4a9","Type":"ContainerDied","Data":"81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00"} Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.529192 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.529198 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a2049067a5414b5746d6ea7c62534722cc178badf04c9dfa7a3a6a562dfa00" Oct 11 03:30:03 crc kubenswrapper[4754]: I1011 03:30:03.595401 4754 scope.go:117] "RemoveContainer" containerID="68e2b4bfdbc154c5daa794c1907bd149283a38ae4af4937f2a848867c26895a4" Oct 11 03:30:04 crc kubenswrapper[4754]: I1011 03:30:04.538413 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerStarted","Data":"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd"} Oct 11 03:30:04 crc kubenswrapper[4754]: I1011 03:30:04.562064 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5h898" podStartSLOduration=4.060346802 podStartE2EDuration="6.561943084s" podCreationTimestamp="2025-10-11 03:29:58 +0000 UTC" firstStartedPulling="2025-10-11 03:30:00.495802459 +0000 UTC m=+1448.054747244" lastFinishedPulling="2025-10-11 03:30:02.997398741 +0000 UTC m=+1450.556343526" observedRunningTime="2025-10-11 03:30:04.553275188 +0000 UTC m=+1452.112219983" watchObservedRunningTime="2025-10-11 03:30:04.561943084 +0000 UTC m=+1452.120887919" Oct 11 03:30:09 crc kubenswrapper[4754]: I1011 03:30:09.432034 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:09 crc kubenswrapper[4754]: I1011 03:30:09.432557 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:09 crc kubenswrapper[4754]: I1011 03:30:09.478027 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:09 crc kubenswrapper[4754]: I1011 03:30:09.617093 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:09 crc kubenswrapper[4754]: I1011 03:30:09.709389 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:30:11 crc kubenswrapper[4754]: I1011 03:30:11.592834 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5h898" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="registry-server" containerID="cri-o://2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd" gracePeriod=2 Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.098592 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.173859 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content\") pod \"51b7695f-0528-49f7-acfe-83fe2577cb04\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.174020 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities\") pod \"51b7695f-0528-49f7-acfe-83fe2577cb04\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.174105 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmd9n\" (UniqueName: \"kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n\") pod \"51b7695f-0528-49f7-acfe-83fe2577cb04\" (UID: \"51b7695f-0528-49f7-acfe-83fe2577cb04\") " Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.174808 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities" (OuterVolumeSpecName: "utilities") pod "51b7695f-0528-49f7-acfe-83fe2577cb04" (UID: "51b7695f-0528-49f7-acfe-83fe2577cb04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.181147 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n" (OuterVolumeSpecName: "kube-api-access-zmd9n") pod "51b7695f-0528-49f7-acfe-83fe2577cb04" (UID: "51b7695f-0528-49f7-acfe-83fe2577cb04"). InnerVolumeSpecName "kube-api-access-zmd9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.220775 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51b7695f-0528-49f7-acfe-83fe2577cb04" (UID: "51b7695f-0528-49f7-acfe-83fe2577cb04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.276165 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.276206 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51b7695f-0528-49f7-acfe-83fe2577cb04-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.276219 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmd9n\" (UniqueName: \"kubernetes.io/projected/51b7695f-0528-49f7-acfe-83fe2577cb04-kube-api-access-zmd9n\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.602985 4754 generic.go:334] "Generic (PLEG): container finished" podID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerID="2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd" exitCode=0 Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.603028 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerDied","Data":"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd"} Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.603048 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5h898" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.603056 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5h898" event={"ID":"51b7695f-0528-49f7-acfe-83fe2577cb04","Type":"ContainerDied","Data":"1042252d6e48e8941044bae7802e0da75e497bf29b74f371b793463f6aed1b6e"} Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.603075 4754 scope.go:117] "RemoveContainer" containerID="2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.627396 4754 scope.go:117] "RemoveContainer" containerID="9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.640246 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.647606 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5h898"] Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.671492 4754 scope.go:117] "RemoveContainer" containerID="563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.706288 4754 scope.go:117] "RemoveContainer" containerID="2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd" Oct 11 03:30:12 crc kubenswrapper[4754]: E1011 03:30:12.706885 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd\": container with ID starting with 2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd not found: ID does not exist" containerID="2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.706927 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd"} err="failed to get container status \"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd\": rpc error: code = NotFound desc = could not find container \"2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd\": container with ID starting with 2e06c517e8658973edaae952b3c85bbb0ae436ea036f5412fae3b36bda3a30bd not found: ID does not exist" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.706952 4754 scope.go:117] "RemoveContainer" containerID="9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96" Oct 11 03:30:12 crc kubenswrapper[4754]: E1011 03:30:12.707324 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96\": container with ID starting with 9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96 not found: ID does not exist" containerID="9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.707374 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96"} err="failed to get container status \"9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96\": rpc error: code = NotFound desc = could not find container \"9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96\": container with ID starting with 9db13f5472d54d48143cee099cc583612b02a7fac3a9b78e6f6e7de44783bc96 not found: ID does not exist" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.707408 4754 scope.go:117] "RemoveContainer" containerID="563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69" Oct 11 03:30:12 crc kubenswrapper[4754]: E1011 03:30:12.707734 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69\": container with ID starting with 563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69 not found: ID does not exist" containerID="563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69" Oct 11 03:30:12 crc kubenswrapper[4754]: I1011 03:30:12.707768 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69"} err="failed to get container status \"563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69\": rpc error: code = NotFound desc = could not find container \"563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69\": container with ID starting with 563d65c2c6eaeec79e338072ede6113648b506fa7a7c2abc32db916774274a69 not found: ID does not exist" Oct 11 03:30:13 crc kubenswrapper[4754]: I1011 03:30:13.097621 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" path="/var/lib/kubelet/pods/51b7695f-0528-49f7-acfe-83fe2577cb04/volumes" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.150180 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:30 crc kubenswrapper[4754]: E1011 03:30:30.151192 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="extract-content" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151205 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="extract-content" Oct 11 03:30:30 crc kubenswrapper[4754]: E1011 03:30:30.151231 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d842ba1d-77ef-4397-a138-849800aff4a9" containerName="collect-profiles" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151237 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d842ba1d-77ef-4397-a138-849800aff4a9" containerName="collect-profiles" Oct 11 03:30:30 crc kubenswrapper[4754]: E1011 03:30:30.151269 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="registry-server" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151275 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="registry-server" Oct 11 03:30:30 crc kubenswrapper[4754]: E1011 03:30:30.151286 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="extract-utilities" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151292 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="extract-utilities" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151445 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d842ba1d-77ef-4397-a138-849800aff4a9" containerName="collect-profiles" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.151461 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b7695f-0528-49f7-acfe-83fe2577cb04" containerName="registry-server" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.153168 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.167202 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.235578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.235718 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.235741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9r66\" (UniqueName: \"kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.337481 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.337859 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9r66\" (UniqueName: \"kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.338027 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.338124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.338488 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.358002 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9r66\" (UniqueName: \"kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66\") pod \"redhat-operators-rsxjk\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.512287 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:30 crc kubenswrapper[4754]: I1011 03:30:30.952016 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:31 crc kubenswrapper[4754]: I1011 03:30:31.795627 4754 generic.go:334] "Generic (PLEG): container finished" podID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerID="b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d" exitCode=0 Oct 11 03:30:31 crc kubenswrapper[4754]: I1011 03:30:31.795866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerDied","Data":"b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d"} Oct 11 03:30:31 crc kubenswrapper[4754]: I1011 03:30:31.796065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerStarted","Data":"4de8415a3e58094fecd9468b950421d1333ec335ee9b003c91de7430758042fa"} Oct 11 03:30:33 crc kubenswrapper[4754]: I1011 03:30:33.818872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerStarted","Data":"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219"} Oct 11 03:30:35 crc kubenswrapper[4754]: I1011 03:30:35.840160 4754 generic.go:334] "Generic (PLEG): container finished" podID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerID="839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219" exitCode=0 Oct 11 03:30:35 crc kubenswrapper[4754]: I1011 03:30:35.840538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerDied","Data":"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219"} Oct 11 03:30:36 crc kubenswrapper[4754]: I1011 03:30:36.849811 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerStarted","Data":"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2"} Oct 11 03:30:36 crc kubenswrapper[4754]: I1011 03:30:36.874055 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rsxjk" podStartSLOduration=2.296178788 podStartE2EDuration="6.874037615s" podCreationTimestamp="2025-10-11 03:30:30 +0000 UTC" firstStartedPulling="2025-10-11 03:30:31.798272966 +0000 UTC m=+1479.357217751" lastFinishedPulling="2025-10-11 03:30:36.376131793 +0000 UTC m=+1483.935076578" observedRunningTime="2025-10-11 03:30:36.871216035 +0000 UTC m=+1484.430160810" watchObservedRunningTime="2025-10-11 03:30:36.874037615 +0000 UTC m=+1484.432982400" Oct 11 03:30:40 crc kubenswrapper[4754]: I1011 03:30:40.513127 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:40 crc kubenswrapper[4754]: I1011 03:30:40.513213 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:41 crc kubenswrapper[4754]: I1011 03:30:41.578181 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rsxjk" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="registry-server" probeResult="failure" output=< Oct 11 03:30:41 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 11 03:30:41 crc kubenswrapper[4754]: > Oct 11 03:30:46 crc kubenswrapper[4754]: I1011 03:30:46.934225 4754 generic.go:334] "Generic (PLEG): container finished" podID="3f2bd066-1dae-406a-8a53-28f620db3fdf" containerID="2aa328f4587bd88700843fbaa8054dd6d9e54a35447c252f6e51c1c2e3e47af3" exitCode=0 Oct 11 03:30:46 crc kubenswrapper[4754]: I1011 03:30:46.934736 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" event={"ID":"3f2bd066-1dae-406a-8a53-28f620db3fdf","Type":"ContainerDied","Data":"2aa328f4587bd88700843fbaa8054dd6d9e54a35447c252f6e51c1c2e3e47af3"} Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.288461 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.352188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key\") pod \"3f2bd066-1dae-406a-8a53-28f620db3fdf\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.352323 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p28qd\" (UniqueName: \"kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd\") pod \"3f2bd066-1dae-406a-8a53-28f620db3fdf\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.352511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory\") pod \"3f2bd066-1dae-406a-8a53-28f620db3fdf\" (UID: \"3f2bd066-1dae-406a-8a53-28f620db3fdf\") " Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.357637 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd" (OuterVolumeSpecName: "kube-api-access-p28qd") pod "3f2bd066-1dae-406a-8a53-28f620db3fdf" (UID: "3f2bd066-1dae-406a-8a53-28f620db3fdf"). InnerVolumeSpecName "kube-api-access-p28qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.376864 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f2bd066-1dae-406a-8a53-28f620db3fdf" (UID: "3f2bd066-1dae-406a-8a53-28f620db3fdf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.382619 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory" (OuterVolumeSpecName: "inventory") pod "3f2bd066-1dae-406a-8a53-28f620db3fdf" (UID: "3f2bd066-1dae-406a-8a53-28f620db3fdf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.454359 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.454387 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2bd066-1dae-406a-8a53-28f620db3fdf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.454397 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p28qd\" (UniqueName: \"kubernetes.io/projected/3f2bd066-1dae-406a-8a53-28f620db3fdf-kube-api-access-p28qd\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.950912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" event={"ID":"3f2bd066-1dae-406a-8a53-28f620db3fdf","Type":"ContainerDied","Data":"0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45"} Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.950953 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a455ed00f1c2fa0635c6eec8f76b0b8e74924a45e08d33f0c64440c7e78ea45" Oct 11 03:30:48 crc kubenswrapper[4754]: I1011 03:30:48.950968 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.036600 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc"] Oct 11 03:30:49 crc kubenswrapper[4754]: E1011 03:30:49.036958 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2bd066-1dae-406a-8a53-28f620db3fdf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.036988 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2bd066-1dae-406a-8a53-28f620db3fdf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.037173 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2bd066-1dae-406a-8a53-28f620db3fdf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.037756 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.039628 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.040162 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.040203 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.040723 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.063247 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc"] Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.066415 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrkrg\" (UniqueName: \"kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.066509 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.066839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.169490 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.170368 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.170558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrkrg\" (UniqueName: \"kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.175871 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.175871 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.188728 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrkrg\" (UniqueName: \"kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pphfc\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.355230 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.847534 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc"] Oct 11 03:30:49 crc kubenswrapper[4754]: I1011 03:30:49.969170 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" event={"ID":"d8ddd70d-d70c-45b4-9ae0-c98c2b701133","Type":"ContainerStarted","Data":"124dd694a012a31d7878f1e958ea116a9eeda2a694f99df72993d98ea5bf02a7"} Oct 11 03:30:50 crc kubenswrapper[4754]: I1011 03:30:50.565518 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:50 crc kubenswrapper[4754]: I1011 03:30:50.615006 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:50 crc kubenswrapper[4754]: I1011 03:30:50.809049 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:50 crc kubenswrapper[4754]: I1011 03:30:50.977923 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" event={"ID":"d8ddd70d-d70c-45b4-9ae0-c98c2b701133","Type":"ContainerStarted","Data":"ebdf342e3cd2a5dffbe51bfbc9cf50219d50df9229ad500372000eaec56c2df7"} Oct 11 03:30:51 crc kubenswrapper[4754]: I1011 03:30:51.003424 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" podStartSLOduration=1.550808768 podStartE2EDuration="2.003404344s" podCreationTimestamp="2025-10-11 03:30:49 +0000 UTC" firstStartedPulling="2025-10-11 03:30:49.852912183 +0000 UTC m=+1497.411856968" lastFinishedPulling="2025-10-11 03:30:50.305507719 +0000 UTC m=+1497.864452544" observedRunningTime="2025-10-11 03:30:50.995500929 +0000 UTC m=+1498.554445714" watchObservedRunningTime="2025-10-11 03:30:51.003404344 +0000 UTC m=+1498.562349129" Oct 11 03:30:51 crc kubenswrapper[4754]: I1011 03:30:51.985125 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rsxjk" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="registry-server" containerID="cri-o://4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2" gracePeriod=2 Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.372465 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.445876 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities\") pod \"7885eed7-e24b-42c7-82be-a010aaf5c288\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.445952 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content\") pod \"7885eed7-e24b-42c7-82be-a010aaf5c288\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.446044 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9r66\" (UniqueName: \"kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66\") pod \"7885eed7-e24b-42c7-82be-a010aaf5c288\" (UID: \"7885eed7-e24b-42c7-82be-a010aaf5c288\") " Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.447131 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities" (OuterVolumeSpecName: "utilities") pod "7885eed7-e24b-42c7-82be-a010aaf5c288" (UID: "7885eed7-e24b-42c7-82be-a010aaf5c288"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.453217 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66" (OuterVolumeSpecName: "kube-api-access-h9r66") pod "7885eed7-e24b-42c7-82be-a010aaf5c288" (UID: "7885eed7-e24b-42c7-82be-a010aaf5c288"). InnerVolumeSpecName "kube-api-access-h9r66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.523509 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7885eed7-e24b-42c7-82be-a010aaf5c288" (UID: "7885eed7-e24b-42c7-82be-a010aaf5c288"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.547734 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.547763 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7885eed7-e24b-42c7-82be-a010aaf5c288-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.547775 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9r66\" (UniqueName: \"kubernetes.io/projected/7885eed7-e24b-42c7-82be-a010aaf5c288-kube-api-access-h9r66\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.996631 4754 generic.go:334] "Generic (PLEG): container finished" podID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerID="4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2" exitCode=0 Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.996695 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsxjk" Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.996717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerDied","Data":"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2"} Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.996783 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsxjk" event={"ID":"7885eed7-e24b-42c7-82be-a010aaf5c288","Type":"ContainerDied","Data":"4de8415a3e58094fecd9468b950421d1333ec335ee9b003c91de7430758042fa"} Oct 11 03:30:52 crc kubenswrapper[4754]: I1011 03:30:52.996806 4754 scope.go:117] "RemoveContainer" containerID="4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.022472 4754 scope.go:117] "RemoveContainer" containerID="839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.030353 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.045197 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rsxjk"] Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.074757 4754 scope.go:117] "RemoveContainer" containerID="b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.115448 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" path="/var/lib/kubelet/pods/7885eed7-e24b-42c7-82be-a010aaf5c288/volumes" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.133431 4754 scope.go:117] "RemoveContainer" containerID="4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2" Oct 11 03:30:53 crc kubenswrapper[4754]: E1011 03:30:53.134465 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2\": container with ID starting with 4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2 not found: ID does not exist" containerID="4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.134517 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2"} err="failed to get container status \"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2\": rpc error: code = NotFound desc = could not find container \"4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2\": container with ID starting with 4a7bcdae6bae50a8da3476edaeb941d5dfecb11fccd387a6df5b7824be2700c2 not found: ID does not exist" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.134556 4754 scope.go:117] "RemoveContainer" containerID="839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219" Oct 11 03:30:53 crc kubenswrapper[4754]: E1011 03:30:53.135051 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219\": container with ID starting with 839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219 not found: ID does not exist" containerID="839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.135077 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219"} err="failed to get container status \"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219\": rpc error: code = NotFound desc = could not find container \"839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219\": container with ID starting with 839448922fee5425508830c88b775498505b2d1b7cd572e902866b3730906219 not found: ID does not exist" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.135098 4754 scope.go:117] "RemoveContainer" containerID="b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d" Oct 11 03:30:53 crc kubenswrapper[4754]: E1011 03:30:53.135550 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d\": container with ID starting with b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d not found: ID does not exist" containerID="b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d" Oct 11 03:30:53 crc kubenswrapper[4754]: I1011 03:30:53.135582 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d"} err="failed to get container status \"b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d\": rpc error: code = NotFound desc = could not find container \"b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d\": container with ID starting with b489e15c4dc7c217203e04b0a1f39505dc6b61e2a3c66708577c1fb5dbc6bf8d not found: ID does not exist" Oct 11 03:30:54 crc kubenswrapper[4754]: I1011 03:30:54.065659 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-cwzw8"] Oct 11 03:30:54 crc kubenswrapper[4754]: I1011 03:30:54.074771 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-cwzw8"] Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.024919 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-s8bp9"] Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.033138 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vmtkc"] Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.041537 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vmtkc"] Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.049010 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-s8bp9"] Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.094083 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="043dd66c-f38e-4d7b-9310-123a0a052ff0" path="/var/lib/kubelet/pods/043dd66c-f38e-4d7b-9310-123a0a052ff0/volumes" Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.094629 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d24e41ae-1434-46b7-a52b-f435f48bc68e" path="/var/lib/kubelet/pods/d24e41ae-1434-46b7-a52b-f435f48bc68e/volumes" Oct 11 03:30:55 crc kubenswrapper[4754]: I1011 03:30:55.095207 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37ec154-6b1a-4a5b-b622-63a4795383b7" path="/var/lib/kubelet/pods/e37ec154-6b1a-4a5b-b622-63a4795383b7/volumes" Oct 11 03:30:56 crc kubenswrapper[4754]: I1011 03:30:56.022635 4754 generic.go:334] "Generic (PLEG): container finished" podID="d8ddd70d-d70c-45b4-9ae0-c98c2b701133" containerID="ebdf342e3cd2a5dffbe51bfbc9cf50219d50df9229ad500372000eaec56c2df7" exitCode=0 Oct 11 03:30:56 crc kubenswrapper[4754]: I1011 03:30:56.022717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" event={"ID":"d8ddd70d-d70c-45b4-9ae0-c98c2b701133","Type":"ContainerDied","Data":"ebdf342e3cd2a5dffbe51bfbc9cf50219d50df9229ad500372000eaec56c2df7"} Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.376212 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.531007 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrkrg\" (UniqueName: \"kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg\") pod \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.531065 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key\") pod \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.531177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory\") pod \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\" (UID: \"d8ddd70d-d70c-45b4-9ae0-c98c2b701133\") " Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.536511 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg" (OuterVolumeSpecName: "kube-api-access-nrkrg") pod "d8ddd70d-d70c-45b4-9ae0-c98c2b701133" (UID: "d8ddd70d-d70c-45b4-9ae0-c98c2b701133"). InnerVolumeSpecName "kube-api-access-nrkrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.562168 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8ddd70d-d70c-45b4-9ae0-c98c2b701133" (UID: "d8ddd70d-d70c-45b4-9ae0-c98c2b701133"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.576387 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory" (OuterVolumeSpecName: "inventory") pod "d8ddd70d-d70c-45b4-9ae0-c98c2b701133" (UID: "d8ddd70d-d70c-45b4-9ae0-c98c2b701133"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.633393 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.633445 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrkrg\" (UniqueName: \"kubernetes.io/projected/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-kube-api-access-nrkrg\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:57 crc kubenswrapper[4754]: I1011 03:30:57.633468 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ddd70d-d70c-45b4-9ae0-c98c2b701133-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.040922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" event={"ID":"d8ddd70d-d70c-45b4-9ae0-c98c2b701133","Type":"ContainerDied","Data":"124dd694a012a31d7878f1e958ea116a9eeda2a694f99df72993d98ea5bf02a7"} Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.040992 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="124dd694a012a31d7878f1e958ea116a9eeda2a694f99df72993d98ea5bf02a7" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.041044 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.136471 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96"] Oct 11 03:30:58 crc kubenswrapper[4754]: E1011 03:30:58.137286 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="extract-content" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137307 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="extract-content" Oct 11 03:30:58 crc kubenswrapper[4754]: E1011 03:30:58.137321 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="registry-server" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137331 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="registry-server" Oct 11 03:30:58 crc kubenswrapper[4754]: E1011 03:30:58.137387 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ddd70d-d70c-45b4-9ae0-c98c2b701133" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137399 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ddd70d-d70c-45b4-9ae0-c98c2b701133" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:58 crc kubenswrapper[4754]: E1011 03:30:58.137431 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="extract-utilities" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137440 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="extract-utilities" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137922 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8ddd70d-d70c-45b4-9ae0-c98c2b701133" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.137988 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7885eed7-e24b-42c7-82be-a010aaf5c288" containerName="registry-server" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.139073 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.143682 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.144070 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.144276 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.144427 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.154121 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96"] Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.244837 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.244917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.245091 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.346826 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.347332 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.347527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.352178 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.352414 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.365297 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5wq96\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.468241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:30:58 crc kubenswrapper[4754]: I1011 03:30:58.997918 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96"] Oct 11 03:30:59 crc kubenswrapper[4754]: I1011 03:30:59.050169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" event={"ID":"80fa82c3-2e3b-4257-b938-099ea565b4a8","Type":"ContainerStarted","Data":"02366bd67b32ac7b2c5ea1fe2babbab3fb98722de612351632504fe8877f42f9"} Oct 11 03:31:00 crc kubenswrapper[4754]: I1011 03:31:00.058183 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" event={"ID":"80fa82c3-2e3b-4257-b938-099ea565b4a8","Type":"ContainerStarted","Data":"3373c52bcce9b94b418a62f2641b544389b051a4ba4613c0a7c94648f5aff9f4"} Oct 11 03:31:00 crc kubenswrapper[4754]: I1011 03:31:00.095400 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" podStartSLOduration=1.6696418469999998 podStartE2EDuration="2.095372552s" podCreationTimestamp="2025-10-11 03:30:58 +0000 UTC" firstStartedPulling="2025-10-11 03:30:59.009305448 +0000 UTC m=+1506.568250253" lastFinishedPulling="2025-10-11 03:30:59.435036173 +0000 UTC m=+1506.993980958" observedRunningTime="2025-10-11 03:31:00.082096005 +0000 UTC m=+1507.641040830" watchObservedRunningTime="2025-10-11 03:31:00.095372552 +0000 UTC m=+1507.654317377" Oct 11 03:31:03 crc kubenswrapper[4754]: I1011 03:31:03.650995 4754 scope.go:117] "RemoveContainer" containerID="0a0fa83857cb4fff543e37f9eb15f650c5ee77e6ebebb020e2c40dbffd363916" Oct 11 03:31:03 crc kubenswrapper[4754]: I1011 03:31:03.679617 4754 scope.go:117] "RemoveContainer" containerID="6c2c1e6e9ec78b54748c60ebc2bf7604121f2f4a89a04b6902475f1e4de9e17c" Oct 11 03:31:03 crc kubenswrapper[4754]: I1011 03:31:03.713144 4754 scope.go:117] "RemoveContainer" containerID="bf06492276a312fd0f33f83bed5c07b9c8c4268a79f0d284a5143d86a68c5bf6" Oct 11 03:31:04 crc kubenswrapper[4754]: I1011 03:31:04.027503 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2706-account-create-gcctp"] Oct 11 03:31:04 crc kubenswrapper[4754]: I1011 03:31:04.038228 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2706-account-create-gcctp"] Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.030539 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-44df-account-create-4fm6g"] Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.039867 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ac5e-account-create-8ph9k"] Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.046815 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-44df-account-create-4fm6g"] Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.053526 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ac5e-account-create-8ph9k"] Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.095655 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bb448c7-4547-4717-9638-49ebda21b7c0" path="/var/lib/kubelet/pods/3bb448c7-4547-4717-9638-49ebda21b7c0/volumes" Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.096178 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b62291-9634-48d3-8d23-6d12aa4c46a3" path="/var/lib/kubelet/pods/68b62291-9634-48d3-8d23-6d12aa4c46a3/volumes" Oct 11 03:31:05 crc kubenswrapper[4754]: I1011 03:31:05.096688 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ea6a3b-0e79-464b-bff3-67a33acca115" path="/var/lib/kubelet/pods/f1ea6a3b-0e79-464b-bff3-67a33acca115/volumes" Oct 11 03:31:26 crc kubenswrapper[4754]: I1011 03:31:26.034875 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-mrg6d"] Oct 11 03:31:26 crc kubenswrapper[4754]: I1011 03:31:26.042902 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-mrg6d"] Oct 11 03:31:26 crc kubenswrapper[4754]: I1011 03:31:26.050406 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6cct7"] Oct 11 03:31:26 crc kubenswrapper[4754]: I1011 03:31:26.057482 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6cct7"] Oct 11 03:31:27 crc kubenswrapper[4754]: I1011 03:31:27.022219 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dw5l4"] Oct 11 03:31:27 crc kubenswrapper[4754]: I1011 03:31:27.028937 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dw5l4"] Oct 11 03:31:27 crc kubenswrapper[4754]: I1011 03:31:27.094250 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce57e33-81b8-4db0-aa86-cb68353f1562" path="/var/lib/kubelet/pods/4ce57e33-81b8-4db0-aa86-cb68353f1562/volumes" Oct 11 03:31:27 crc kubenswrapper[4754]: I1011 03:31:27.094769 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed0ab40-9674-4778-aaa4-8dc24d13f10a" path="/var/lib/kubelet/pods/6ed0ab40-9674-4778-aaa4-8dc24d13f10a/volumes" Oct 11 03:31:27 crc kubenswrapper[4754]: I1011 03:31:27.095257 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae" path="/var/lib/kubelet/pods/a6ec089c-1d8c-49ae-96ca-0bf5f36f1aae/volumes" Oct 11 03:31:28 crc kubenswrapper[4754]: I1011 03:31:28.031702 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-zt4j7"] Oct 11 03:31:28 crc kubenswrapper[4754]: I1011 03:31:28.042999 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-zt4j7"] Oct 11 03:31:29 crc kubenswrapper[4754]: I1011 03:31:29.102012 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd76760-5f2d-4c44-91f7-c8b2277df563" path="/var/lib/kubelet/pods/3dd76760-5f2d-4c44-91f7-c8b2277df563/volumes" Oct 11 03:31:31 crc kubenswrapper[4754]: I1011 03:31:31.037131 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zjs8v"] Oct 11 03:31:31 crc kubenswrapper[4754]: I1011 03:31:31.046845 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zjs8v"] Oct 11 03:31:31 crc kubenswrapper[4754]: I1011 03:31:31.096987 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0073f3a4-5ab0-4ee7-9793-08f96385ffd0" path="/var/lib/kubelet/pods/0073f3a4-5ab0-4ee7-9793-08f96385ffd0/volumes" Oct 11 03:31:34 crc kubenswrapper[4754]: I1011 03:31:34.032886 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5f8f-account-create-7h2dd"] Oct 11 03:31:34 crc kubenswrapper[4754]: I1011 03:31:34.047159 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5f8f-account-create-7h2dd"] Oct 11 03:31:35 crc kubenswrapper[4754]: I1011 03:31:35.097710 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0007db7e-72b2-4e2a-b6b2-5cd94b44badf" path="/var/lib/kubelet/pods/0007db7e-72b2-4e2a-b6b2-5cd94b44badf/volumes" Oct 11 03:31:40 crc kubenswrapper[4754]: I1011 03:31:40.461850 4754 generic.go:334] "Generic (PLEG): container finished" podID="80fa82c3-2e3b-4257-b938-099ea565b4a8" containerID="3373c52bcce9b94b418a62f2641b544389b051a4ba4613c0a7c94648f5aff9f4" exitCode=0 Oct 11 03:31:40 crc kubenswrapper[4754]: I1011 03:31:40.461952 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" event={"ID":"80fa82c3-2e3b-4257-b938-099ea565b4a8","Type":"ContainerDied","Data":"3373c52bcce9b94b418a62f2641b544389b051a4ba4613c0a7c94648f5aff9f4"} Oct 11 03:31:41 crc kubenswrapper[4754]: I1011 03:31:41.956952 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.062431 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9\") pod \"80fa82c3-2e3b-4257-b938-099ea565b4a8\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.062517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory\") pod \"80fa82c3-2e3b-4257-b938-099ea565b4a8\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.062539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key\") pod \"80fa82c3-2e3b-4257-b938-099ea565b4a8\" (UID: \"80fa82c3-2e3b-4257-b938-099ea565b4a8\") " Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.068406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9" (OuterVolumeSpecName: "kube-api-access-nhqw9") pod "80fa82c3-2e3b-4257-b938-099ea565b4a8" (UID: "80fa82c3-2e3b-4257-b938-099ea565b4a8"). InnerVolumeSpecName "kube-api-access-nhqw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.089887 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80fa82c3-2e3b-4257-b938-099ea565b4a8" (UID: "80fa82c3-2e3b-4257-b938-099ea565b4a8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.093678 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory" (OuterVolumeSpecName: "inventory") pod "80fa82c3-2e3b-4257-b938-099ea565b4a8" (UID: "80fa82c3-2e3b-4257-b938-099ea565b4a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.164368 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/80fa82c3-2e3b-4257-b938-099ea565b4a8-kube-api-access-nhqw9\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.164408 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.164416 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80fa82c3-2e3b-4257-b938-099ea565b4a8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.483571 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" event={"ID":"80fa82c3-2e3b-4257-b938-099ea565b4a8","Type":"ContainerDied","Data":"02366bd67b32ac7b2c5ea1fe2babbab3fb98722de612351632504fe8877f42f9"} Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.483615 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02366bd67b32ac7b2c5ea1fe2babbab3fb98722de612351632504fe8877f42f9" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.483622 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.555782 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl"] Oct 11 03:31:42 crc kubenswrapper[4754]: E1011 03:31:42.556432 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fa82c3-2e3b-4257-b938-099ea565b4a8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.556450 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fa82c3-2e3b-4257-b938-099ea565b4a8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.556608 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fa82c3-2e3b-4257-b938-099ea565b4a8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.557184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.558896 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.559186 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.559251 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.559495 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.578310 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl"] Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.675280 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgfpg\" (UniqueName: \"kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.675363 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.675423 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.776744 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgfpg\" (UniqueName: \"kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.777196 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.777337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.781596 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.781726 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.813711 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgfpg\" (UniqueName: \"kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:42 crc kubenswrapper[4754]: I1011 03:31:42.882045 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:43 crc kubenswrapper[4754]: I1011 03:31:43.444022 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl"] Oct 11 03:31:43 crc kubenswrapper[4754]: I1011 03:31:43.453618 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:31:43 crc kubenswrapper[4754]: I1011 03:31:43.501004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" event={"ID":"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0","Type":"ContainerStarted","Data":"78ad660855c0005fd0df67028035c8674d1aac8e6f873e424e8da7913184f679"} Oct 11 03:31:44 crc kubenswrapper[4754]: I1011 03:31:44.515618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" event={"ID":"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0","Type":"ContainerStarted","Data":"86d45a8c693ee631c57dc120264ddac224be3a4c3ef232a13a2ff22b6fcbfc26"} Oct 11 03:31:44 crc kubenswrapper[4754]: I1011 03:31:44.544578 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" podStartSLOduration=2.111795171 podStartE2EDuration="2.544526604s" podCreationTimestamp="2025-10-11 03:31:42 +0000 UTC" firstStartedPulling="2025-10-11 03:31:43.453323695 +0000 UTC m=+1551.012268500" lastFinishedPulling="2025-10-11 03:31:43.886055128 +0000 UTC m=+1551.444999933" observedRunningTime="2025-10-11 03:31:44.535353734 +0000 UTC m=+1552.094298539" watchObservedRunningTime="2025-10-11 03:31:44.544526604 +0000 UTC m=+1552.103471399" Oct 11 03:31:46 crc kubenswrapper[4754]: I1011 03:31:46.030290 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-336f-account-create-v5pnq"] Oct 11 03:31:46 crc kubenswrapper[4754]: I1011 03:31:46.040150 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-336f-account-create-v5pnq"] Oct 11 03:31:47 crc kubenswrapper[4754]: I1011 03:31:47.045070 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-453c-account-create-kxp6w"] Oct 11 03:31:47 crc kubenswrapper[4754]: I1011 03:31:47.055609 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-453c-account-create-kxp6w"] Oct 11 03:31:47 crc kubenswrapper[4754]: I1011 03:31:47.093895 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b929a42-00e5-4d36-a15e-15b2678adb2a" path="/var/lib/kubelet/pods/4b929a42-00e5-4d36-a15e-15b2678adb2a/volumes" Oct 11 03:31:47 crc kubenswrapper[4754]: I1011 03:31:47.094415 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0101651-749d-42ea-8355-9e82199d2fe5" path="/var/lib/kubelet/pods/f0101651-749d-42ea-8355-9e82199d2fe5/volumes" Oct 11 03:31:48 crc kubenswrapper[4754]: I1011 03:31:48.553043 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" containerID="86d45a8c693ee631c57dc120264ddac224be3a4c3ef232a13a2ff22b6fcbfc26" exitCode=0 Oct 11 03:31:48 crc kubenswrapper[4754]: I1011 03:31:48.553102 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" event={"ID":"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0","Type":"ContainerDied","Data":"86d45a8c693ee631c57dc120264ddac224be3a4c3ef232a13a2ff22b6fcbfc26"} Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.024617 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gd655"] Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.032542 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gd655"] Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.048727 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.129976 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key\") pod \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.130158 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgfpg\" (UniqueName: \"kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg\") pod \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.130230 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory\") pod \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\" (UID: \"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0\") " Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.136318 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg" (OuterVolumeSpecName: "kube-api-access-lgfpg") pod "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" (UID: "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0"). InnerVolumeSpecName "kube-api-access-lgfpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.159724 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory" (OuterVolumeSpecName: "inventory") pod "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" (UID: "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.172879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" (UID: "f6ed9c9b-8831-4e8b-87a3-705c9ca217e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.232398 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgfpg\" (UniqueName: \"kubernetes.io/projected/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-kube-api-access-lgfpg\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.232439 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.232454 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.576458 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" event={"ID":"f6ed9c9b-8831-4e8b-87a3-705c9ca217e0","Type":"ContainerDied","Data":"78ad660855c0005fd0df67028035c8674d1aac8e6f873e424e8da7913184f679"} Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.576501 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78ad660855c0005fd0df67028035c8674d1aac8e6f873e424e8da7913184f679" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.576561 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.674441 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6"] Oct 11 03:31:50 crc kubenswrapper[4754]: E1011 03:31:50.675141 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.675159 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.675451 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.676168 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6"] Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.676250 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.693784 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.694070 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.694108 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.694249 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.749178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6ngl\" (UniqueName: \"kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.749414 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.749587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.851002 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.851113 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6ngl\" (UniqueName: \"kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.851158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.858780 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.859287 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:50 crc kubenswrapper[4754]: I1011 03:31:50.873780 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6ngl\" (UniqueName: \"kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:51 crc kubenswrapper[4754]: I1011 03:31:51.014198 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:31:51 crc kubenswrapper[4754]: I1011 03:31:51.103035 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2224e5c6-2dfc-4e51-b2ee-19cc25084573" path="/var/lib/kubelet/pods/2224e5c6-2dfc-4e51-b2ee-19cc25084573/volumes" Oct 11 03:31:51 crc kubenswrapper[4754]: I1011 03:31:51.341247 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6"] Oct 11 03:31:51 crc kubenswrapper[4754]: I1011 03:31:51.586348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" event={"ID":"ecaafa15-e064-4e7b-af73-e7c8d8be6323","Type":"ContainerStarted","Data":"820758fa89435010d608c401fb326740ccc3fb4cb095d20c565f7543ef601029"} Oct 11 03:31:52 crc kubenswrapper[4754]: I1011 03:31:52.597941 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" event={"ID":"ecaafa15-e064-4e7b-af73-e7c8d8be6323","Type":"ContainerStarted","Data":"1de2bdd3cf5b5ee9cba517b90419cc4e3f3acc1a35b9867faebe39f8ca3dd2da"} Oct 11 03:31:52 crc kubenswrapper[4754]: I1011 03:31:52.620522 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" podStartSLOduration=2.20774433 podStartE2EDuration="2.620506457s" podCreationTimestamp="2025-10-11 03:31:50 +0000 UTC" firstStartedPulling="2025-10-11 03:31:51.359341037 +0000 UTC m=+1558.918285822" lastFinishedPulling="2025-10-11 03:31:51.772103164 +0000 UTC m=+1559.331047949" observedRunningTime="2025-10-11 03:31:52.61500045 +0000 UTC m=+1560.173945245" watchObservedRunningTime="2025-10-11 03:31:52.620506457 +0000 UTC m=+1560.179451242" Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.040936 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xxmz8"] Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.051291 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xxmz8"] Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.093027 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17eed71a-41c2-4333-b346-3156a8b6aaee" path="/var/lib/kubelet/pods/17eed71a-41c2-4333-b346-3156a8b6aaee/volumes" Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.829571 4754 scope.go:117] "RemoveContainer" containerID="74fdff6fd203561129ec1e5505fd464048df93295666e98d4658fdfd116b07ce" Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.885130 4754 scope.go:117] "RemoveContainer" containerID="465587eae06745a7342f9df4c5fdc149065c88c75cd088df91808ba607d2ce7f" Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.949467 4754 scope.go:117] "RemoveContainer" containerID="23b0b42896893b37a6d3ebf8931703b08187ac6495fab267cd1dae908145f7cf" Oct 11 03:32:03 crc kubenswrapper[4754]: I1011 03:32:03.990800 4754 scope.go:117] "RemoveContainer" containerID="5eb48ba0203ae5ff77ebd9dacd996dd0fae0667b22ac438d5bce4d75346d16f9" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.021514 4754 scope.go:117] "RemoveContainer" containerID="5fcb9911414b0c21743b75a7d8084bfcac5d1ec5aa7a3070be2cd0ef4c1f37fe" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.063689 4754 scope.go:117] "RemoveContainer" containerID="c67b92a36f1cda6583bf21d08ef40b6f604290b8f75215f26f73c91974694c3a" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.097776 4754 scope.go:117] "RemoveContainer" containerID="c20ad0f1d594d79351874f09f89b7a9380d075d05cabe162e83f3bee0ed7943c" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.126746 4754 scope.go:117] "RemoveContainer" containerID="fa14052dff70eeb8ea5fc5ce7b7df7bd6af2a71fb293e179dcee756f474d9494" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.172138 4754 scope.go:117] "RemoveContainer" containerID="d4d16df0900c21a0dc62f10b161e1ea20529f278f2175e359da129ac3e0e751a" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.198530 4754 scope.go:117] "RemoveContainer" containerID="c6efcfb156d574d02111ffa339bb781cdaace237c752f7c8aa386e3be55d81f4" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.243615 4754 scope.go:117] "RemoveContainer" containerID="4ab2dce136d937ad284c4bf18488c37c197409308a2461e1cdcdb8ce1d692f8d" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.265881 4754 scope.go:117] "RemoveContainer" containerID="928cdb0d0dcecc112a271b1cb9fc84b2f66347567fbfdacf4e3970c36ada1033" Oct 11 03:32:04 crc kubenswrapper[4754]: I1011 03:32:04.300539 4754 scope.go:117] "RemoveContainer" containerID="6609bdad63bd73ee2a1b36a7c111f004159ff5b59def53f9597a71049b2f118f" Oct 11 03:32:17 crc kubenswrapper[4754]: I1011 03:32:17.036514 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-jmzkc"] Oct 11 03:32:17 crc kubenswrapper[4754]: I1011 03:32:17.048719 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-jmzkc"] Oct 11 03:32:17 crc kubenswrapper[4754]: I1011 03:32:17.092685 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93df894a-1420-42c8-9352-5acd38935208" path="/var/lib/kubelet/pods/93df894a-1420-42c8-9352-5acd38935208/volumes" Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.030357 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vqm6s"] Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.047440 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-h7c6l"] Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.056956 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-h7c6l"] Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.064340 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vqm6s"] Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.104490 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ed9167b-ca5f-4fe5-b362-2d2bdef58d30" path="/var/lib/kubelet/pods/3ed9167b-ca5f-4fe5-b362-2d2bdef58d30/volumes" Oct 11 03:32:25 crc kubenswrapper[4754]: I1011 03:32:25.106242 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78bf9306-4051-4919-9276-e89b67330f04" path="/var/lib/kubelet/pods/78bf9306-4051-4919-9276-e89b67330f04/volumes" Oct 11 03:32:30 crc kubenswrapper[4754]: I1011 03:32:30.737034 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:32:30 crc kubenswrapper[4754]: I1011 03:32:30.737700 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.065107 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rwhqn"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.075072 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-j2hk9"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.103334 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mkp24"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.106568 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rwhqn"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.116879 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mkp24"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.123249 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-j2hk9"] Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.157663 4754 generic.go:334] "Generic (PLEG): container finished" podID="ecaafa15-e064-4e7b-af73-e7c8d8be6323" containerID="1de2bdd3cf5b5ee9cba517b90419cc4e3f3acc1a35b9867faebe39f8ca3dd2da" exitCode=2 Oct 11 03:32:53 crc kubenswrapper[4754]: I1011 03:32:53.157705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" event={"ID":"ecaafa15-e064-4e7b-af73-e7c8d8be6323","Type":"ContainerDied","Data":"1de2bdd3cf5b5ee9cba517b90419cc4e3f3acc1a35b9867faebe39f8ca3dd2da"} Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.586144 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.686885 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory\") pod \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.687059 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key\") pod \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.687216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6ngl\" (UniqueName: \"kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl\") pod \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\" (UID: \"ecaafa15-e064-4e7b-af73-e7c8d8be6323\") " Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.696235 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl" (OuterVolumeSpecName: "kube-api-access-n6ngl") pod "ecaafa15-e064-4e7b-af73-e7c8d8be6323" (UID: "ecaafa15-e064-4e7b-af73-e7c8d8be6323"). InnerVolumeSpecName "kube-api-access-n6ngl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.715622 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory" (OuterVolumeSpecName: "inventory") pod "ecaafa15-e064-4e7b-af73-e7c8d8be6323" (UID: "ecaafa15-e064-4e7b-af73-e7c8d8be6323"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.733549 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecaafa15-e064-4e7b-af73-e7c8d8be6323" (UID: "ecaafa15-e064-4e7b-af73-e7c8d8be6323"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.790245 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6ngl\" (UniqueName: \"kubernetes.io/projected/ecaafa15-e064-4e7b-af73-e7c8d8be6323-kube-api-access-n6ngl\") on node \"crc\" DevicePath \"\"" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.790299 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:32:54 crc kubenswrapper[4754]: I1011 03:32:54.790319 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecaafa15-e064-4e7b-af73-e7c8d8be6323-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.099744 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a5329e2-6d7f-477c-9496-347d8a49f4df" path="/var/lib/kubelet/pods/3a5329e2-6d7f-477c-9496-347d8a49f4df/volumes" Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.101394 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2d4813-8e67-4e79-8aad-5519f6355ced" path="/var/lib/kubelet/pods/8f2d4813-8e67-4e79-8aad-5519f6355ced/volumes" Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.102622 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3625e97-9933-4bf5-9253-48a0465301fd" path="/var/lib/kubelet/pods/b3625e97-9933-4bf5-9253-48a0465301fd/volumes" Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.176606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" event={"ID":"ecaafa15-e064-4e7b-af73-e7c8d8be6323","Type":"ContainerDied","Data":"820758fa89435010d608c401fb326740ccc3fb4cb095d20c565f7543ef601029"} Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.176945 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="820758fa89435010d608c401fb326740ccc3fb4cb095d20c565f7543ef601029" Oct 11 03:32:55 crc kubenswrapper[4754]: I1011 03:32:55.176821 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6" Oct 11 03:33:00 crc kubenswrapper[4754]: I1011 03:33:00.736781 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:33:00 crc kubenswrapper[4754]: I1011 03:33:00.737696 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:33:01 crc kubenswrapper[4754]: I1011 03:33:01.033011 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d868-account-create-6gct8"] Oct 11 03:33:01 crc kubenswrapper[4754]: I1011 03:33:01.044615 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d868-account-create-6gct8"] Oct 11 03:33:01 crc kubenswrapper[4754]: I1011 03:33:01.096755 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9fc5ae-cabb-483c-8aec-d0b5b478d54b" path="/var/lib/kubelet/pods/ea9fc5ae-cabb-483c-8aec-d0b5b478d54b/volumes" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.034240 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98"] Oct 11 03:33:02 crc kubenswrapper[4754]: E1011 03:33:02.034579 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecaafa15-e064-4e7b-af73-e7c8d8be6323" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.034593 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecaafa15-e064-4e7b-af73-e7c8d8be6323" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.034781 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecaafa15-e064-4e7b-af73-e7c8d8be6323" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.035509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.038549 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.039446 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.039511 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.039592 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.065793 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98"] Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.141143 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.141238 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.141326 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp66r\" (UniqueName: \"kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.243570 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.243663 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp66r\" (UniqueName: \"kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.243805 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.249488 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.251248 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.267283 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp66r\" (UniqueName: \"kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ksk98\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.373222 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:02 crc kubenswrapper[4754]: I1011 03:33:02.883415 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98"] Oct 11 03:33:02 crc kubenswrapper[4754]: W1011 03:33:02.893356 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d77e345_4dff_45d5_8d70_ab42d2106305.slice/crio-8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878 WatchSource:0}: Error finding container 8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878: Status 404 returned error can't find the container with id 8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878 Oct 11 03:33:03 crc kubenswrapper[4754]: I1011 03:33:03.280122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" event={"ID":"5d77e345-4dff-45d5-8d70-ab42d2106305","Type":"ContainerStarted","Data":"8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878"} Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.302231 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" event={"ID":"5d77e345-4dff-45d5-8d70-ab42d2106305","Type":"ContainerStarted","Data":"16e79848db9701330717df0fc7c855a3b4fd5ed424f36ab44b49bfb33233e310"} Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.324228 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" podStartSLOduration=1.90685811 podStartE2EDuration="2.324205637s" podCreationTimestamp="2025-10-11 03:33:02 +0000 UTC" firstStartedPulling="2025-10-11 03:33:02.895050895 +0000 UTC m=+1630.453995680" lastFinishedPulling="2025-10-11 03:33:03.312398412 +0000 UTC m=+1630.871343207" observedRunningTime="2025-10-11 03:33:04.321715086 +0000 UTC m=+1631.880659941" watchObservedRunningTime="2025-10-11 03:33:04.324205637 +0000 UTC m=+1631.883150422" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.514304 4754 scope.go:117] "RemoveContainer" containerID="8f88f642291bfe492294f35c68e1d2403be9a8bde3b27e3c9b7bc4a4476747e4" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.538761 4754 scope.go:117] "RemoveContainer" containerID="bef964d39dce76ff5239b9ab3edc3c38ae1a78e707cacdf50fcdd413b2f8b188" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.592318 4754 scope.go:117] "RemoveContainer" containerID="e5ff870bad281ae0edbce039f90a1579be77878a17536f1a8f6d5d254ae2f0db" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.619779 4754 scope.go:117] "RemoveContainer" containerID="a4f0973a5edbcd3d0e55781498362458aed82b78c14fbc2fb36987ee44beb8dc" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.653610 4754 scope.go:117] "RemoveContainer" containerID="89781785b76dbf12528636ee8c09f05422be24e2f63fa4e5b664991c246401e2" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.703667 4754 scope.go:117] "RemoveContainer" containerID="a3b0d8781811253873f50d0a476fcd7e6e914deed9e39ea0d34ae12bac2a316c" Oct 11 03:33:04 crc kubenswrapper[4754]: I1011 03:33:04.743116 4754 scope.go:117] "RemoveContainer" containerID="129073840aa223733a780a4be753edfc279bc404440474759f083d597fd029df" Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.026717 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-bc5d-account-create-qt2kw"] Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.034442 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-6de7-account-create-d2fp2"] Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.040895 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-bc5d-account-create-qt2kw"] Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.047162 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-6de7-account-create-d2fp2"] Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.097740 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c7c15bb-cace-4445-8e0c-d07fe698b3a2" path="/var/lib/kubelet/pods/1c7c15bb-cace-4445-8e0c-d07fe698b3a2/volumes" Oct 11 03:33:11 crc kubenswrapper[4754]: I1011 03:33:11.098447 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80dd29fb-8f91-4375-97f8-70c9fff47f7f" path="/var/lib/kubelet/pods/80dd29fb-8f91-4375-97f8-70c9fff47f7f/volumes" Oct 11 03:33:30 crc kubenswrapper[4754]: I1011 03:33:30.736408 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:33:30 crc kubenswrapper[4754]: I1011 03:33:30.736922 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:33:30 crc kubenswrapper[4754]: I1011 03:33:30.736987 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:33:30 crc kubenswrapper[4754]: I1011 03:33:30.737658 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:33:30 crc kubenswrapper[4754]: I1011 03:33:30.737711 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" gracePeriod=600 Oct 11 03:33:30 crc kubenswrapper[4754]: E1011 03:33:30.861679 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:33:31 crc kubenswrapper[4754]: I1011 03:33:31.530296 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" exitCode=0 Oct 11 03:33:31 crc kubenswrapper[4754]: I1011 03:33:31.530350 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5"} Oct 11 03:33:31 crc kubenswrapper[4754]: I1011 03:33:31.530395 4754 scope.go:117] "RemoveContainer" containerID="0a990d667503e08ae23916f0d1f2284aa036850e8ad400ec6a7c3e5e103a1cf6" Oct 11 03:33:31 crc kubenswrapper[4754]: I1011 03:33:31.530941 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:33:31 crc kubenswrapper[4754]: E1011 03:33:31.531219 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:33:32 crc kubenswrapper[4754]: I1011 03:33:32.044623 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2grg7"] Oct 11 03:33:32 crc kubenswrapper[4754]: I1011 03:33:32.052446 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2grg7"] Oct 11 03:33:33 crc kubenswrapper[4754]: I1011 03:33:33.117656 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5708fd0-491c-4117-a569-8c183034a478" path="/var/lib/kubelet/pods/a5708fd0-491c-4117-a569-8c183034a478/volumes" Oct 11 03:33:46 crc kubenswrapper[4754]: I1011 03:33:46.082989 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:33:46 crc kubenswrapper[4754]: E1011 03:33:46.083840 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:33:52 crc kubenswrapper[4754]: I1011 03:33:52.718805 4754 generic.go:334] "Generic (PLEG): container finished" podID="5d77e345-4dff-45d5-8d70-ab42d2106305" containerID="16e79848db9701330717df0fc7c855a3b4fd5ed424f36ab44b49bfb33233e310" exitCode=0 Oct 11 03:33:52 crc kubenswrapper[4754]: I1011 03:33:52.718862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" event={"ID":"5d77e345-4dff-45d5-8d70-ab42d2106305","Type":"ContainerDied","Data":"16e79848db9701330717df0fc7c855a3b4fd5ed424f36ab44b49bfb33233e310"} Oct 11 03:33:53 crc kubenswrapper[4754]: I1011 03:33:53.039064 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bsdg6"] Oct 11 03:33:53 crc kubenswrapper[4754]: I1011 03:33:53.051254 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bsdg6"] Oct 11 03:33:53 crc kubenswrapper[4754]: I1011 03:33:53.107979 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185f8b44-dd99-48d0-940a-7f14690d6728" path="/var/lib/kubelet/pods/185f8b44-dd99-48d0-940a-7f14690d6728/volumes" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.115922 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.267636 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key\") pod \"5d77e345-4dff-45d5-8d70-ab42d2106305\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.267716 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") pod \"5d77e345-4dff-45d5-8d70-ab42d2106305\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.267892 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp66r\" (UniqueName: \"kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r\") pod \"5d77e345-4dff-45d5-8d70-ab42d2106305\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.279661 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r" (OuterVolumeSpecName: "kube-api-access-pp66r") pod "5d77e345-4dff-45d5-8d70-ab42d2106305" (UID: "5d77e345-4dff-45d5-8d70-ab42d2106305"). InnerVolumeSpecName "kube-api-access-pp66r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:33:54 crc kubenswrapper[4754]: E1011 03:33:54.301297 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory podName:5d77e345-4dff-45d5-8d70-ab42d2106305 nodeName:}" failed. No retries permitted until 2025-10-11 03:33:54.801266257 +0000 UTC m=+1682.360211052 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory") pod "5d77e345-4dff-45d5-8d70-ab42d2106305" (UID: "5d77e345-4dff-45d5-8d70-ab42d2106305") : error deleting /var/lib/kubelet/pods/5d77e345-4dff-45d5-8d70-ab42d2106305/volume-subpaths: remove /var/lib/kubelet/pods/5d77e345-4dff-45d5-8d70-ab42d2106305/volume-subpaths: no such file or directory Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.304166 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d77e345-4dff-45d5-8d70-ab42d2106305" (UID: "5d77e345-4dff-45d5-8d70-ab42d2106305"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.369911 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp66r\" (UniqueName: \"kubernetes.io/projected/5d77e345-4dff-45d5-8d70-ab42d2106305-kube-api-access-pp66r\") on node \"crc\" DevicePath \"\"" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.369938 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.744272 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" event={"ID":"5d77e345-4dff-45d5-8d70-ab42d2106305","Type":"ContainerDied","Data":"8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878"} Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.744586 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8057185b1f97b0dee670694cbdbe38a056dc7b608a59f5c8578414899cfe1878" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.744383 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.823995 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hflrd"] Oct 11 03:33:54 crc kubenswrapper[4754]: E1011 03:33:54.824447 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d77e345-4dff-45d5-8d70-ab42d2106305" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.824475 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d77e345-4dff-45d5-8d70-ab42d2106305" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.824706 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d77e345-4dff-45d5-8d70-ab42d2106305" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.825472 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.835857 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hflrd"] Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.879450 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") pod \"5d77e345-4dff-45d5-8d70-ab42d2106305\" (UID: \"5d77e345-4dff-45d5-8d70-ab42d2106305\") " Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.883760 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory" (OuterVolumeSpecName: "inventory") pod "5d77e345-4dff-45d5-8d70-ab42d2106305" (UID: "5d77e345-4dff-45d5-8d70-ab42d2106305"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.981152 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.981356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.981412 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rpg9\" (UniqueName: \"kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:54 crc kubenswrapper[4754]: I1011 03:33:54.981520 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d77e345-4dff-45d5-8d70-ab42d2106305-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.033073 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zn2qx"] Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.045220 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zn2qx"] Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.082745 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.082910 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.082958 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rpg9\" (UniqueName: \"kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.092886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.093123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.098341 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1639fb4-c6c2-48b4-a13c-6f5166229dac" path="/var/lib/kubelet/pods/a1639fb4-c6c2-48b4-a13c-6f5166229dac/volumes" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.106990 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rpg9\" (UniqueName: \"kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9\") pod \"ssh-known-hosts-edpm-deployment-hflrd\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.176670 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.705695 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hflrd"] Oct 11 03:33:55 crc kubenswrapper[4754]: I1011 03:33:55.754052 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" event={"ID":"35535b44-8b92-4908-ac94-f761fa8bc9ac","Type":"ContainerStarted","Data":"c0e5051b0b9daf5ef13d00d479982e9fdee8a796a42715f815198c44ab95aa85"} Oct 11 03:33:56 crc kubenswrapper[4754]: I1011 03:33:56.765661 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" event={"ID":"35535b44-8b92-4908-ac94-f761fa8bc9ac","Type":"ContainerStarted","Data":"ae2c3c0cda800ced07e9b4dc1ea0a208ab09b0ce05cd32f8259d7268254642d0"} Oct 11 03:33:56 crc kubenswrapper[4754]: I1011 03:33:56.789210 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" podStartSLOduration=2.23505934 podStartE2EDuration="2.789190053s" podCreationTimestamp="2025-10-11 03:33:54 +0000 UTC" firstStartedPulling="2025-10-11 03:33:55.713673988 +0000 UTC m=+1683.272618773" lastFinishedPulling="2025-10-11 03:33:56.267804711 +0000 UTC m=+1683.826749486" observedRunningTime="2025-10-11 03:33:56.788656968 +0000 UTC m=+1684.347601753" watchObservedRunningTime="2025-10-11 03:33:56.789190053 +0000 UTC m=+1684.348134838" Oct 11 03:33:59 crc kubenswrapper[4754]: I1011 03:33:59.085377 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:33:59 crc kubenswrapper[4754]: E1011 03:33:59.086011 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:34:03 crc kubenswrapper[4754]: I1011 03:34:03.828069 4754 generic.go:334] "Generic (PLEG): container finished" podID="35535b44-8b92-4908-ac94-f761fa8bc9ac" containerID="ae2c3c0cda800ced07e9b4dc1ea0a208ab09b0ce05cd32f8259d7268254642d0" exitCode=0 Oct 11 03:34:03 crc kubenswrapper[4754]: I1011 03:34:03.828150 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" event={"ID":"35535b44-8b92-4908-ac94-f761fa8bc9ac","Type":"ContainerDied","Data":"ae2c3c0cda800ced07e9b4dc1ea0a208ab09b0ce05cd32f8259d7268254642d0"} Oct 11 03:34:04 crc kubenswrapper[4754]: I1011 03:34:04.851736 4754 scope.go:117] "RemoveContainer" containerID="173034bae9024d38dc086dac8fe19051ee21d16c172abd6e7f17b1f0d75f1e4d" Oct 11 03:34:04 crc kubenswrapper[4754]: I1011 03:34:04.909446 4754 scope.go:117] "RemoveContainer" containerID="85a54549196e899d44baab37dac0be4bb16c869a7593d0408d0aaacdea31dcb8" Oct 11 03:34:04 crc kubenswrapper[4754]: I1011 03:34:04.982979 4754 scope.go:117] "RemoveContainer" containerID="c3fe91e98bcfa01925eb2916edbbae71a896560ec239041be0266feb1631fe6c" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.008684 4754 scope.go:117] "RemoveContainer" containerID="4caeb350ec2b7d9b8a3111468651e4cee18a21ea1846dc2d8151b23ad0f9c2b7" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.047496 4754 scope.go:117] "RemoveContainer" containerID="bf61717581adc95c2a00a63f6c90644fce0059e268b08e9d7f9a2489429a64fe" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.160128 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.273192 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam\") pod \"35535b44-8b92-4908-ac94-f761fa8bc9ac\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.273583 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0\") pod \"35535b44-8b92-4908-ac94-f761fa8bc9ac\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.273700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rpg9\" (UniqueName: \"kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9\") pod \"35535b44-8b92-4908-ac94-f761fa8bc9ac\" (UID: \"35535b44-8b92-4908-ac94-f761fa8bc9ac\") " Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.277902 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9" (OuterVolumeSpecName: "kube-api-access-6rpg9") pod "35535b44-8b92-4908-ac94-f761fa8bc9ac" (UID: "35535b44-8b92-4908-ac94-f761fa8bc9ac"). InnerVolumeSpecName "kube-api-access-6rpg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.298757 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "35535b44-8b92-4908-ac94-f761fa8bc9ac" (UID: "35535b44-8b92-4908-ac94-f761fa8bc9ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.298922 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "35535b44-8b92-4908-ac94-f761fa8bc9ac" (UID: "35535b44-8b92-4908-ac94-f761fa8bc9ac"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.376314 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.376360 4754 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/35535b44-8b92-4908-ac94-f761fa8bc9ac-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.376375 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rpg9\" (UniqueName: \"kubernetes.io/projected/35535b44-8b92-4908-ac94-f761fa8bc9ac-kube-api-access-6rpg9\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.849827 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" event={"ID":"35535b44-8b92-4908-ac94-f761fa8bc9ac","Type":"ContainerDied","Data":"c0e5051b0b9daf5ef13d00d479982e9fdee8a796a42715f815198c44ab95aa85"} Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.849902 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0e5051b0b9daf5ef13d00d479982e9fdee8a796a42715f815198c44ab95aa85" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.849989 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hflrd" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.947156 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7"] Oct 11 03:34:05 crc kubenswrapper[4754]: E1011 03:34:05.947592 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35535b44-8b92-4908-ac94-f761fa8bc9ac" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.947608 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35535b44-8b92-4908-ac94-f761fa8bc9ac" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.947808 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="35535b44-8b92-4908-ac94-f761fa8bc9ac" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.948406 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.952063 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.952313 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.952586 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.952930 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:34:05 crc kubenswrapper[4754]: I1011 03:34:05.970652 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7"] Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.087880 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.087932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzg9z\" (UniqueName: \"kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.088069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.189382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzg9z\" (UniqueName: \"kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.189709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.190063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.195045 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.195803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.206209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzg9z\" (UniqueName: \"kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vqzb7\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.266875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.782685 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7"] Oct 11 03:34:06 crc kubenswrapper[4754]: I1011 03:34:06.857976 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" event={"ID":"16965691-4e28-447e-b770-4a49c5206327","Type":"ContainerStarted","Data":"890319255b2adc1b4ac349eccc096259641508917e066d4d4ca4fb389407f906"} Oct 11 03:34:07 crc kubenswrapper[4754]: I1011 03:34:07.868094 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" event={"ID":"16965691-4e28-447e-b770-4a49c5206327","Type":"ContainerStarted","Data":"dd1600c607668d9198bdc0e6b7e4b1931902b073c8c10381a0e819436d295504"} Oct 11 03:34:07 crc kubenswrapper[4754]: I1011 03:34:07.888290 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" podStartSLOduration=2.445124075 podStartE2EDuration="2.888271078s" podCreationTimestamp="2025-10-11 03:34:05 +0000 UTC" firstStartedPulling="2025-10-11 03:34:06.791254757 +0000 UTC m=+1694.350199542" lastFinishedPulling="2025-10-11 03:34:07.23440176 +0000 UTC m=+1694.793346545" observedRunningTime="2025-10-11 03:34:07.884296887 +0000 UTC m=+1695.443241682" watchObservedRunningTime="2025-10-11 03:34:07.888271078 +0000 UTC m=+1695.447215863" Oct 11 03:34:10 crc kubenswrapper[4754]: I1011 03:34:10.084420 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:34:10 crc kubenswrapper[4754]: E1011 03:34:10.086315 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:34:15 crc kubenswrapper[4754]: I1011 03:34:15.937353 4754 generic.go:334] "Generic (PLEG): container finished" podID="16965691-4e28-447e-b770-4a49c5206327" containerID="dd1600c607668d9198bdc0e6b7e4b1931902b073c8c10381a0e819436d295504" exitCode=0 Oct 11 03:34:15 crc kubenswrapper[4754]: I1011 03:34:15.937450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" event={"ID":"16965691-4e28-447e-b770-4a49c5206327","Type":"ContainerDied","Data":"dd1600c607668d9198bdc0e6b7e4b1931902b073c8c10381a0e819436d295504"} Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.405282 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.522653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory\") pod \"16965691-4e28-447e-b770-4a49c5206327\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.522786 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzg9z\" (UniqueName: \"kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z\") pod \"16965691-4e28-447e-b770-4a49c5206327\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.522814 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key\") pod \"16965691-4e28-447e-b770-4a49c5206327\" (UID: \"16965691-4e28-447e-b770-4a49c5206327\") " Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.536200 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z" (OuterVolumeSpecName: "kube-api-access-zzg9z") pod "16965691-4e28-447e-b770-4a49c5206327" (UID: "16965691-4e28-447e-b770-4a49c5206327"). InnerVolumeSpecName "kube-api-access-zzg9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.550380 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory" (OuterVolumeSpecName: "inventory") pod "16965691-4e28-447e-b770-4a49c5206327" (UID: "16965691-4e28-447e-b770-4a49c5206327"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.565149 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "16965691-4e28-447e-b770-4a49c5206327" (UID: "16965691-4e28-447e-b770-4a49c5206327"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.624450 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.624498 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzg9z\" (UniqueName: \"kubernetes.io/projected/16965691-4e28-447e-b770-4a49c5206327-kube-api-access-zzg9z\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.624513 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16965691-4e28-447e-b770-4a49c5206327-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.956897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" event={"ID":"16965691-4e28-447e-b770-4a49c5206327","Type":"ContainerDied","Data":"890319255b2adc1b4ac349eccc096259641508917e066d4d4ca4fb389407f906"} Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.956929 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7" Oct 11 03:34:17 crc kubenswrapper[4754]: I1011 03:34:17.956940 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="890319255b2adc1b4ac349eccc096259641508917e066d4d4ca4fb389407f906" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.036408 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z"] Oct 11 03:34:18 crc kubenswrapper[4754]: E1011 03:34:18.037078 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16965691-4e28-447e-b770-4a49c5206327" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.037182 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="16965691-4e28-447e-b770-4a49c5206327" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.037427 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="16965691-4e28-447e-b770-4a49c5206327" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.039047 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.041835 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.042257 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.042565 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.043345 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.051376 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z"] Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.134994 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.135133 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.135219 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hftmx\" (UniqueName: \"kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.236538 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.236636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.236698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hftmx\" (UniqueName: \"kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.241865 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.242614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.251352 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hftmx\" (UniqueName: \"kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.361985 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.826678 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z"] Oct 11 03:34:18 crc kubenswrapper[4754]: I1011 03:34:18.964493 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" event={"ID":"fbc2e287-2c45-46c0-8872-8540353c52bc","Type":"ContainerStarted","Data":"64de1e84fe366f4ed5f549f8d0c8f5259a03be5724b00960e685c91e2ff41bd3"} Oct 11 03:34:19 crc kubenswrapper[4754]: I1011 03:34:19.972771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" event={"ID":"fbc2e287-2c45-46c0-8872-8540353c52bc","Type":"ContainerStarted","Data":"e4376dc3e7a4540db08e27e856b82300ccffc59453e78805b2264ce66a5cbd96"} Oct 11 03:34:19 crc kubenswrapper[4754]: I1011 03:34:19.989989 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" podStartSLOduration=1.607946306 podStartE2EDuration="1.989952079s" podCreationTimestamp="2025-10-11 03:34:18 +0000 UTC" firstStartedPulling="2025-10-11 03:34:18.831447238 +0000 UTC m=+1706.390392023" lastFinishedPulling="2025-10-11 03:34:19.213453011 +0000 UTC m=+1706.772397796" observedRunningTime="2025-10-11 03:34:19.98926646 +0000 UTC m=+1707.548211275" watchObservedRunningTime="2025-10-11 03:34:19.989952079 +0000 UTC m=+1707.548896884" Oct 11 03:34:25 crc kubenswrapper[4754]: I1011 03:34:25.083529 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:34:25 crc kubenswrapper[4754]: E1011 03:34:25.084096 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:34:30 crc kubenswrapper[4754]: I1011 03:34:30.089392 4754 generic.go:334] "Generic (PLEG): container finished" podID="fbc2e287-2c45-46c0-8872-8540353c52bc" containerID="e4376dc3e7a4540db08e27e856b82300ccffc59453e78805b2264ce66a5cbd96" exitCode=0 Oct 11 03:34:30 crc kubenswrapper[4754]: I1011 03:34:30.089564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" event={"ID":"fbc2e287-2c45-46c0-8872-8540353c52bc","Type":"ContainerDied","Data":"e4376dc3e7a4540db08e27e856b82300ccffc59453e78805b2264ce66a5cbd96"} Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.581322 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.717305 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hftmx\" (UniqueName: \"kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx\") pod \"fbc2e287-2c45-46c0-8872-8540353c52bc\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.718177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory\") pod \"fbc2e287-2c45-46c0-8872-8540353c52bc\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.718302 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key\") pod \"fbc2e287-2c45-46c0-8872-8540353c52bc\" (UID: \"fbc2e287-2c45-46c0-8872-8540353c52bc\") " Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.722474 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx" (OuterVolumeSpecName: "kube-api-access-hftmx") pod "fbc2e287-2c45-46c0-8872-8540353c52bc" (UID: "fbc2e287-2c45-46c0-8872-8540353c52bc"). InnerVolumeSpecName "kube-api-access-hftmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.743870 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fbc2e287-2c45-46c0-8872-8540353c52bc" (UID: "fbc2e287-2c45-46c0-8872-8540353c52bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.745550 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory" (OuterVolumeSpecName: "inventory") pod "fbc2e287-2c45-46c0-8872-8540353c52bc" (UID: "fbc2e287-2c45-46c0-8872-8540353c52bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.820290 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.820326 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hftmx\" (UniqueName: \"kubernetes.io/projected/fbc2e287-2c45-46c0-8872-8540353c52bc-kube-api-access-hftmx\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:31 crc kubenswrapper[4754]: I1011 03:34:31.820342 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbc2e287-2c45-46c0-8872-8540353c52bc-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:34:32 crc kubenswrapper[4754]: I1011 03:34:32.108588 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" event={"ID":"fbc2e287-2c45-46c0-8872-8540353c52bc","Type":"ContainerDied","Data":"64de1e84fe366f4ed5f549f8d0c8f5259a03be5724b00960e685c91e2ff41bd3"} Oct 11 03:34:32 crc kubenswrapper[4754]: I1011 03:34:32.108644 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64de1e84fe366f4ed5f549f8d0c8f5259a03be5724b00960e685c91e2ff41bd3" Oct 11 03:34:32 crc kubenswrapper[4754]: I1011 03:34:32.108687 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z" Oct 11 03:34:39 crc kubenswrapper[4754]: I1011 03:34:39.083598 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:34:39 crc kubenswrapper[4754]: E1011 03:34:39.084294 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:34:40 crc kubenswrapper[4754]: I1011 03:34:40.037814 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mfrtj"] Oct 11 03:34:40 crc kubenswrapper[4754]: I1011 03:34:40.051235 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mfrtj"] Oct 11 03:34:41 crc kubenswrapper[4754]: I1011 03:34:41.105203 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4c309f-66cb-46e5-b6b0-8599a4019cd3" path="/var/lib/kubelet/pods/ab4c309f-66cb-46e5-b6b0-8599a4019cd3/volumes" Oct 11 03:34:53 crc kubenswrapper[4754]: I1011 03:34:53.088758 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:34:53 crc kubenswrapper[4754]: E1011 03:34:53.089446 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:35:05 crc kubenswrapper[4754]: I1011 03:35:05.083669 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:35:05 crc kubenswrapper[4754]: E1011 03:35:05.084418 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:35:05 crc kubenswrapper[4754]: I1011 03:35:05.212387 4754 scope.go:117] "RemoveContainer" containerID="32fe382470935e34e03dadf3080fb4856f160b0a457434596b165c1adf7d06c7" Oct 11 03:35:17 crc kubenswrapper[4754]: I1011 03:35:17.083937 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:35:17 crc kubenswrapper[4754]: E1011 03:35:17.084718 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:35:28 crc kubenswrapper[4754]: I1011 03:35:28.083464 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:35:28 crc kubenswrapper[4754]: E1011 03:35:28.084450 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:35:42 crc kubenswrapper[4754]: I1011 03:35:42.083699 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:35:42 crc kubenswrapper[4754]: E1011 03:35:42.084448 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:35:55 crc kubenswrapper[4754]: I1011 03:35:55.083807 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:35:55 crc kubenswrapper[4754]: E1011 03:35:55.084773 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:36:07 crc kubenswrapper[4754]: I1011 03:36:07.086349 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:36:07 crc kubenswrapper[4754]: E1011 03:36:07.087116 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:36:18 crc kubenswrapper[4754]: I1011 03:36:18.083832 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:36:18 crc kubenswrapper[4754]: E1011 03:36:18.084944 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:36:31 crc kubenswrapper[4754]: I1011 03:36:31.083861 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:36:31 crc kubenswrapper[4754]: E1011 03:36:31.084566 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:36:46 crc kubenswrapper[4754]: I1011 03:36:46.083380 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:36:46 crc kubenswrapper[4754]: E1011 03:36:46.084398 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:36:58 crc kubenswrapper[4754]: I1011 03:36:58.083600 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:36:58 crc kubenswrapper[4754]: E1011 03:36:58.085517 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:37:13 crc kubenswrapper[4754]: I1011 03:37:13.090135 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:37:13 crc kubenswrapper[4754]: E1011 03:37:13.090793 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:37:27 crc kubenswrapper[4754]: I1011 03:37:27.085330 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:37:27 crc kubenswrapper[4754]: E1011 03:37:27.086502 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:37:43 crc kubenswrapper[4754]: I1011 03:37:43.097854 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:37:43 crc kubenswrapper[4754]: E1011 03:37:43.100460 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:37:58 crc kubenswrapper[4754]: I1011 03:37:58.084126 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:37:58 crc kubenswrapper[4754]: E1011 03:37:58.086065 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:38:09 crc kubenswrapper[4754]: I1011 03:38:09.084537 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:38:09 crc kubenswrapper[4754]: E1011 03:38:09.085490 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:38:20 crc kubenswrapper[4754]: I1011 03:38:20.083792 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:38:20 crc kubenswrapper[4754]: E1011 03:38:20.084593 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:38:35 crc kubenswrapper[4754]: I1011 03:38:35.083252 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:38:36 crc kubenswrapper[4754]: I1011 03:38:36.233915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a"} Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.134367 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.145878 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.166594 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.175411 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-t5p9z"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.182279 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.188740 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vkhbl"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.202446 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.209278 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hflrd"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.215727 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pphfc"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.221675 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9m4xk"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.227986 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.234149 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.239658 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q45s6"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.245319 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hflrd"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.251003 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.256163 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.261292 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.268789 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5wq96"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.275765 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hq4j6"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.281734 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-blnfh"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.288448 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vqzb7"] Oct 11 03:39:11 crc kubenswrapper[4754]: I1011 03:39:11.294869 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ksk98"] Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.111542 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16965691-4e28-447e-b770-4a49c5206327" path="/var/lib/kubelet/pods/16965691-4e28-447e-b770-4a49c5206327/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.112849 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35535b44-8b92-4908-ac94-f761fa8bc9ac" path="/var/lib/kubelet/pods/35535b44-8b92-4908-ac94-f761fa8bc9ac/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.113719 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2bd066-1dae-406a-8a53-28f620db3fdf" path="/var/lib/kubelet/pods/3f2bd066-1dae-406a-8a53-28f620db3fdf/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.114760 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d3017f-d07c-4e07-b696-0ce8a6bc8e71" path="/var/lib/kubelet/pods/53d3017f-d07c-4e07-b696-0ce8a6bc8e71/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.115621 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d77e345-4dff-45d5-8d70-ab42d2106305" path="/var/lib/kubelet/pods/5d77e345-4dff-45d5-8d70-ab42d2106305/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.116172 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80fa82c3-2e3b-4257-b938-099ea565b4a8" path="/var/lib/kubelet/pods/80fa82c3-2e3b-4257-b938-099ea565b4a8/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.116712 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8ddd70d-d70c-45b4-9ae0-c98c2b701133" path="/var/lib/kubelet/pods/d8ddd70d-d70c-45b4-9ae0-c98c2b701133/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.117689 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecaafa15-e064-4e7b-af73-e7c8d8be6323" path="/var/lib/kubelet/pods/ecaafa15-e064-4e7b-af73-e7c8d8be6323/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.118249 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f009017c-d229-4243-922f-cd0ded520f61" path="/var/lib/kubelet/pods/f009017c-d229-4243-922f-cd0ded520f61/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.118766 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ed9c9b-8831-4e8b-87a3-705c9ca217e0" path="/var/lib/kubelet/pods/f6ed9c9b-8831-4e8b-87a3-705c9ca217e0/volumes" Oct 11 03:39:13 crc kubenswrapper[4754]: I1011 03:39:13.119690 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbc2e287-2c45-46c0-8872-8540353c52bc" path="/var/lib/kubelet/pods/fbc2e287-2c45-46c0-8872-8540353c52bc/volumes" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.743803 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4"] Oct 11 03:39:16 crc kubenswrapper[4754]: E1011 03:39:16.745906 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc2e287-2c45-46c0-8872-8540353c52bc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.746030 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc2e287-2c45-46c0-8872-8540353c52bc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.746310 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc2e287-2c45-46c0-8872-8540353c52bc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.747028 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.748889 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.749717 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.750227 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.750417 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.753556 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.763011 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4"] Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.823434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmh9\" (UniqueName: \"kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.823495 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.823543 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.823584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.823939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.926025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.926092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.926127 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.926222 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.926269 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmh9\" (UniqueName: \"kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.936822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.937157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.937201 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.940626 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:16 crc kubenswrapper[4754]: I1011 03:39:16.941539 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmh9\" (UniqueName: \"kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:17 crc kubenswrapper[4754]: I1011 03:39:17.082875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:17 crc kubenswrapper[4754]: I1011 03:39:17.611325 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4"] Oct 11 03:39:17 crc kubenswrapper[4754]: I1011 03:39:17.631005 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:39:18 crc kubenswrapper[4754]: I1011 03:39:18.566903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" event={"ID":"11b8f98b-8629-4ebd-b448-89d3056021cf","Type":"ContainerStarted","Data":"5267c3d5712f5fc6eff7ece78b774947339871873ca8d8ce5ffac73506530f07"} Oct 11 03:39:18 crc kubenswrapper[4754]: I1011 03:39:18.567329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" event={"ID":"11b8f98b-8629-4ebd-b448-89d3056021cf","Type":"ContainerStarted","Data":"6cabcee6ba4de5d2ee9c49b600222cabcf2b3aa8850ab3f4ffb4da337b6ea698"} Oct 11 03:39:18 crc kubenswrapper[4754]: I1011 03:39:18.586698 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" podStartSLOduration=2.154627841 podStartE2EDuration="2.5866769s" podCreationTimestamp="2025-10-11 03:39:16 +0000 UTC" firstStartedPulling="2025-10-11 03:39:17.630651321 +0000 UTC m=+2005.189596106" lastFinishedPulling="2025-10-11 03:39:18.06270038 +0000 UTC m=+2005.621645165" observedRunningTime="2025-10-11 03:39:18.581826783 +0000 UTC m=+2006.140771568" watchObservedRunningTime="2025-10-11 03:39:18.5866769 +0000 UTC m=+2006.145621685" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.726355 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.728929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.743371 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.809156 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.809505 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.809628 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqv46\" (UniqueName: \"kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.911413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.911497 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqv46\" (UniqueName: \"kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.911527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.912036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.912072 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:21 crc kubenswrapper[4754]: I1011 03:39:21.937943 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqv46\" (UniqueName: \"kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46\") pod \"certified-operators-tg95n\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:22 crc kubenswrapper[4754]: I1011 03:39:22.051471 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:22 crc kubenswrapper[4754]: I1011 03:39:22.626619 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:22 crc kubenswrapper[4754]: W1011 03:39:22.633326 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded9df8a4_2c1f_48ff_a255_22c806db9260.slice/crio-e75c6e0b7b0b95bb1cd9da6a6fe0af632340fbd424ad48b1783d4b215c5426ec WatchSource:0}: Error finding container e75c6e0b7b0b95bb1cd9da6a6fe0af632340fbd424ad48b1783d4b215c5426ec: Status 404 returned error can't find the container with id e75c6e0b7b0b95bb1cd9da6a6fe0af632340fbd424ad48b1783d4b215c5426ec Oct 11 03:39:23 crc kubenswrapper[4754]: I1011 03:39:23.613590 4754 generic.go:334] "Generic (PLEG): container finished" podID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerID="4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199" exitCode=0 Oct 11 03:39:23 crc kubenswrapper[4754]: I1011 03:39:23.613711 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerDied","Data":"4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199"} Oct 11 03:39:23 crc kubenswrapper[4754]: I1011 03:39:23.614228 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerStarted","Data":"e75c6e0b7b0b95bb1cd9da6a6fe0af632340fbd424ad48b1783d4b215c5426ec"} Oct 11 03:39:24 crc kubenswrapper[4754]: I1011 03:39:24.624020 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerStarted","Data":"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36"} Oct 11 03:39:25 crc kubenswrapper[4754]: I1011 03:39:25.635678 4754 generic.go:334] "Generic (PLEG): container finished" podID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerID="5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36" exitCode=0 Oct 11 03:39:25 crc kubenswrapper[4754]: I1011 03:39:25.635731 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerDied","Data":"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36"} Oct 11 03:39:26 crc kubenswrapper[4754]: I1011 03:39:26.645932 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerStarted","Data":"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826"} Oct 11 03:39:26 crc kubenswrapper[4754]: I1011 03:39:26.671937 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tg95n" podStartSLOduration=3.166324254 podStartE2EDuration="5.671919801s" podCreationTimestamp="2025-10-11 03:39:21 +0000 UTC" firstStartedPulling="2025-10-11 03:39:23.620155976 +0000 UTC m=+2011.179100801" lastFinishedPulling="2025-10-11 03:39:26.125751563 +0000 UTC m=+2013.684696348" observedRunningTime="2025-10-11 03:39:26.664928053 +0000 UTC m=+2014.223872838" watchObservedRunningTime="2025-10-11 03:39:26.671919801 +0000 UTC m=+2014.230864586" Oct 11 03:39:29 crc kubenswrapper[4754]: I1011 03:39:29.674972 4754 generic.go:334] "Generic (PLEG): container finished" podID="11b8f98b-8629-4ebd-b448-89d3056021cf" containerID="5267c3d5712f5fc6eff7ece78b774947339871873ca8d8ce5ffac73506530f07" exitCode=0 Oct 11 03:39:29 crc kubenswrapper[4754]: I1011 03:39:29.675015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" event={"ID":"11b8f98b-8629-4ebd-b448-89d3056021cf","Type":"ContainerDied","Data":"5267c3d5712f5fc6eff7ece78b774947339871873ca8d8ce5ffac73506530f07"} Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.185595 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.297088 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key\") pod \"11b8f98b-8629-4ebd-b448-89d3056021cf\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.297254 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph\") pod \"11b8f98b-8629-4ebd-b448-89d3056021cf\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.297278 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle\") pod \"11b8f98b-8629-4ebd-b448-89d3056021cf\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.297365 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory\") pod \"11b8f98b-8629-4ebd-b448-89d3056021cf\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.297409 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvmh9\" (UniqueName: \"kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9\") pod \"11b8f98b-8629-4ebd-b448-89d3056021cf\" (UID: \"11b8f98b-8629-4ebd-b448-89d3056021cf\") " Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.303516 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph" (OuterVolumeSpecName: "ceph") pod "11b8f98b-8629-4ebd-b448-89d3056021cf" (UID: "11b8f98b-8629-4ebd-b448-89d3056021cf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.303885 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9" (OuterVolumeSpecName: "kube-api-access-cvmh9") pod "11b8f98b-8629-4ebd-b448-89d3056021cf" (UID: "11b8f98b-8629-4ebd-b448-89d3056021cf"). InnerVolumeSpecName "kube-api-access-cvmh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.307187 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "11b8f98b-8629-4ebd-b448-89d3056021cf" (UID: "11b8f98b-8629-4ebd-b448-89d3056021cf"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.323115 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory" (OuterVolumeSpecName: "inventory") pod "11b8f98b-8629-4ebd-b448-89d3056021cf" (UID: "11b8f98b-8629-4ebd-b448-89d3056021cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.328370 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "11b8f98b-8629-4ebd-b448-89d3056021cf" (UID: "11b8f98b-8629-4ebd-b448-89d3056021cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.399269 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.399307 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvmh9\" (UniqueName: \"kubernetes.io/projected/11b8f98b-8629-4ebd-b448-89d3056021cf-kube-api-access-cvmh9\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.399321 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.399360 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.399374 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b8f98b-8629-4ebd-b448-89d3056021cf-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.693716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" event={"ID":"11b8f98b-8629-4ebd-b448-89d3056021cf","Type":"ContainerDied","Data":"6cabcee6ba4de5d2ee9c49b600222cabcf2b3aa8850ab3f4ffb4da337b6ea698"} Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.693934 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cabcee6ba4de5d2ee9c49b600222cabcf2b3aa8850ab3f4ffb4da337b6ea698" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.693790 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.762628 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw"] Oct 11 03:39:31 crc kubenswrapper[4754]: E1011 03:39:31.763249 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b8f98b-8629-4ebd-b448-89d3056021cf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.763342 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b8f98b-8629-4ebd-b448-89d3056021cf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.763643 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b8f98b-8629-4ebd-b448-89d3056021cf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.764448 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.766498 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.766741 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.767411 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.769327 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.770151 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.788662 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw"] Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.907829 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.907915 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs6f5\" (UniqueName: \"kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.907948 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.908024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:31 crc kubenswrapper[4754]: I1011 03:39:31.908059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.010099 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.010167 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.010232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.010295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs6f5\" (UniqueName: \"kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.010330 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.014754 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.015236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.015259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.015528 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.028106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs6f5\" (UniqueName: \"kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.052883 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.052921 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.086008 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.107246 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.606259 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw"] Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.704436 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" event={"ID":"f77837ec-b2bb-4683-bc5f-21ec5145910c","Type":"ContainerStarted","Data":"c22a2e67b22e5d2f0a6e50045e3855c4482d2485182c4d99717890b57dafad51"} Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.754366 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:32 crc kubenswrapper[4754]: I1011 03:39:32.797325 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:33 crc kubenswrapper[4754]: I1011 03:39:33.713343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" event={"ID":"f77837ec-b2bb-4683-bc5f-21ec5145910c","Type":"ContainerStarted","Data":"fd3dd95860e9667b262c73f21c6b98864c7df335f650f7a6bf6f96f09745971b"} Oct 11 03:39:33 crc kubenswrapper[4754]: I1011 03:39:33.735191 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" podStartSLOduration=2.325097006 podStartE2EDuration="2.735171045s" podCreationTimestamp="2025-10-11 03:39:31 +0000 UTC" firstStartedPulling="2025-10-11 03:39:32.624976664 +0000 UTC m=+2020.183921449" lastFinishedPulling="2025-10-11 03:39:33.035050703 +0000 UTC m=+2020.593995488" observedRunningTime="2025-10-11 03:39:33.731062519 +0000 UTC m=+2021.290007314" watchObservedRunningTime="2025-10-11 03:39:33.735171045 +0000 UTC m=+2021.294115840" Oct 11 03:39:34 crc kubenswrapper[4754]: I1011 03:39:34.724570 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tg95n" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="registry-server" containerID="cri-o://fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826" gracePeriod=2 Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.209509 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.370004 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqv46\" (UniqueName: \"kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46\") pod \"ed9df8a4-2c1f-48ff-a255-22c806db9260\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.370187 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content\") pod \"ed9df8a4-2c1f-48ff-a255-22c806db9260\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.370227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities\") pod \"ed9df8a4-2c1f-48ff-a255-22c806db9260\" (UID: \"ed9df8a4-2c1f-48ff-a255-22c806db9260\") " Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.371376 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities" (OuterVolumeSpecName: "utilities") pod "ed9df8a4-2c1f-48ff-a255-22c806db9260" (UID: "ed9df8a4-2c1f-48ff-a255-22c806db9260"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.381231 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46" (OuterVolumeSpecName: "kube-api-access-jqv46") pod "ed9df8a4-2c1f-48ff-a255-22c806db9260" (UID: "ed9df8a4-2c1f-48ff-a255-22c806db9260"). InnerVolumeSpecName "kube-api-access-jqv46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.409828 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed9df8a4-2c1f-48ff-a255-22c806db9260" (UID: "ed9df8a4-2c1f-48ff-a255-22c806db9260"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.472437 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqv46\" (UniqueName: \"kubernetes.io/projected/ed9df8a4-2c1f-48ff-a255-22c806db9260-kube-api-access-jqv46\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.472475 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.472485 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9df8a4-2c1f-48ff-a255-22c806db9260-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.736023 4754 generic.go:334] "Generic (PLEG): container finished" podID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerID="fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826" exitCode=0 Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.736057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerDied","Data":"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826"} Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.736115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg95n" event={"ID":"ed9df8a4-2c1f-48ff-a255-22c806db9260","Type":"ContainerDied","Data":"e75c6e0b7b0b95bb1cd9da6a6fe0af632340fbd424ad48b1783d4b215c5426ec"} Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.736136 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg95n" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.736162 4754 scope.go:117] "RemoveContainer" containerID="fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.759287 4754 scope.go:117] "RemoveContainer" containerID="5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.787579 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.794125 4754 scope.go:117] "RemoveContainer" containerID="4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.796025 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tg95n"] Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.831840 4754 scope.go:117] "RemoveContainer" containerID="fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826" Oct 11 03:39:35 crc kubenswrapper[4754]: E1011 03:39:35.832329 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826\": container with ID starting with fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826 not found: ID does not exist" containerID="fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.832366 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826"} err="failed to get container status \"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826\": rpc error: code = NotFound desc = could not find container \"fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826\": container with ID starting with fff09766b4b9c57180ad48c2824ae0fefb68e298a1de5e00f301e7e184b52826 not found: ID does not exist" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.832386 4754 scope.go:117] "RemoveContainer" containerID="5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36" Oct 11 03:39:35 crc kubenswrapper[4754]: E1011 03:39:35.832649 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36\": container with ID starting with 5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36 not found: ID does not exist" containerID="5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.832672 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36"} err="failed to get container status \"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36\": rpc error: code = NotFound desc = could not find container \"5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36\": container with ID starting with 5ad112c9b1e646dc0316b0a92b73fc5df0a7e1379351f6c18e76482ac3bb4c36 not found: ID does not exist" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.832686 4754 scope.go:117] "RemoveContainer" containerID="4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199" Oct 11 03:39:35 crc kubenswrapper[4754]: E1011 03:39:35.832956 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199\": container with ID starting with 4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199 not found: ID does not exist" containerID="4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199" Oct 11 03:39:35 crc kubenswrapper[4754]: I1011 03:39:35.833046 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199"} err="failed to get container status \"4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199\": rpc error: code = NotFound desc = could not find container \"4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199\": container with ID starting with 4ff5fc6cabebb3a73f41a6fdf2b001e0b735c3f4d3e19208ec081d5ce659a199 not found: ID does not exist" Oct 11 03:39:37 crc kubenswrapper[4754]: I1011 03:39:37.100579 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" path="/var/lib/kubelet/pods/ed9df8a4-2c1f-48ff-a255-22c806db9260/volumes" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.376483 4754 scope.go:117] "RemoveContainer" containerID="16e79848db9701330717df0fc7c855a3b4fd5ed424f36ab44b49bfb33233e310" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.420335 4754 scope.go:117] "RemoveContainer" containerID="1de2bdd3cf5b5ee9cba517b90419cc4e3f3acc1a35b9867faebe39f8ca3dd2da" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.457788 4754 scope.go:117] "RemoveContainer" containerID="3373c52bcce9b94b418a62f2641b544389b051a4ba4613c0a7c94648f5aff9f4" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.496774 4754 scope.go:117] "RemoveContainer" containerID="ebdf342e3cd2a5dffbe51bfbc9cf50219d50df9229ad500372000eaec56c2df7" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.520598 4754 scope.go:117] "RemoveContainer" containerID="f66067f6561342ebf8ff8fffea0fc6e2dc47748a8131518c4db4ddb1f6e241c6" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.588001 4754 scope.go:117] "RemoveContainer" containerID="760df319ae572dd7e0b2d876932b44166a3b0a1e1a5b71f252c89e576bf75357" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.611806 4754 scope.go:117] "RemoveContainer" containerID="ae2c3c0cda800ced07e9b4dc1ea0a208ab09b0ce05cd32f8259d7268254642d0" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.637149 4754 scope.go:117] "RemoveContainer" containerID="2aa328f4587bd88700843fbaa8054dd6d9e54a35447c252f6e51c1c2e3e47af3" Oct 11 03:40:05 crc kubenswrapper[4754]: I1011 03:40:05.673578 4754 scope.go:117] "RemoveContainer" containerID="86d45a8c693ee631c57dc120264ddac224be3a4c3ef232a13a2ff22b6fcbfc26" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.268325 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:31 crc kubenswrapper[4754]: E1011 03:40:31.269193 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="registry-server" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.269206 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="registry-server" Oct 11 03:40:31 crc kubenswrapper[4754]: E1011 03:40:31.269227 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="extract-utilities" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.269234 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="extract-utilities" Oct 11 03:40:31 crc kubenswrapper[4754]: E1011 03:40:31.269251 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="extract-content" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.269257 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="extract-content" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.269424 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9df8a4-2c1f-48ff-a255-22c806db9260" containerName="registry-server" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.270577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.326834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.328607 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.328872 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.329279 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsq7z\" (UniqueName: \"kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.430670 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsq7z\" (UniqueName: \"kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.430748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.430815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.431313 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.431331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.458728 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsq7z\" (UniqueName: \"kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z\") pod \"community-operators-wr9zv\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:31 crc kubenswrapper[4754]: I1011 03:40:31.642906 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:32 crc kubenswrapper[4754]: I1011 03:40:32.194301 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:32 crc kubenswrapper[4754]: I1011 03:40:32.212730 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerStarted","Data":"d02f64f920ee4173a871d8fdef6ef24f3e457544b76a7342ec015140bd894a3e"} Oct 11 03:40:33 crc kubenswrapper[4754]: I1011 03:40:33.226792 4754 generic.go:334] "Generic (PLEG): container finished" podID="a870983f-fa46-479f-be4a-4c9975a83268" containerID="2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14" exitCode=0 Oct 11 03:40:33 crc kubenswrapper[4754]: I1011 03:40:33.227251 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerDied","Data":"2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14"} Oct 11 03:40:34 crc kubenswrapper[4754]: I1011 03:40:34.235192 4754 generic.go:334] "Generic (PLEG): container finished" podID="a870983f-fa46-479f-be4a-4c9975a83268" containerID="a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c" exitCode=0 Oct 11 03:40:34 crc kubenswrapper[4754]: I1011 03:40:34.235397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerDied","Data":"a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c"} Oct 11 03:40:35 crc kubenswrapper[4754]: I1011 03:40:35.245838 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerStarted","Data":"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777"} Oct 11 03:40:35 crc kubenswrapper[4754]: I1011 03:40:35.263247 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wr9zv" podStartSLOduration=2.6567826979999998 podStartE2EDuration="4.263227284s" podCreationTimestamp="2025-10-11 03:40:31 +0000 UTC" firstStartedPulling="2025-10-11 03:40:33.233195137 +0000 UTC m=+2080.792139952" lastFinishedPulling="2025-10-11 03:40:34.839639753 +0000 UTC m=+2082.398584538" observedRunningTime="2025-10-11 03:40:35.261999159 +0000 UTC m=+2082.820943944" watchObservedRunningTime="2025-10-11 03:40:35.263227284 +0000 UTC m=+2082.822172059" Oct 11 03:40:41 crc kubenswrapper[4754]: I1011 03:40:41.643370 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:41 crc kubenswrapper[4754]: I1011 03:40:41.644386 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:41 crc kubenswrapper[4754]: I1011 03:40:41.697502 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:42 crc kubenswrapper[4754]: I1011 03:40:42.367186 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:42 crc kubenswrapper[4754]: I1011 03:40:42.437031 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.323433 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wr9zv" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="registry-server" containerID="cri-o://753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777" gracePeriod=2 Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.354216 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.357734 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.359511 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.359570 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4mjb\" (UniqueName: \"kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.359614 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.366892 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.461088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.461133 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4mjb\" (UniqueName: \"kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.461165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.461674 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.462299 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.491259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4mjb\" (UniqueName: \"kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb\") pod \"redhat-operators-4sq6k\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:44 crc kubenswrapper[4754]: I1011 03:40:44.783206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.026257 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.169676 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities\") pod \"a870983f-fa46-479f-be4a-4c9975a83268\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.170068 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content\") pod \"a870983f-fa46-479f-be4a-4c9975a83268\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.170105 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsq7z\" (UniqueName: \"kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z\") pod \"a870983f-fa46-479f-be4a-4c9975a83268\" (UID: \"a870983f-fa46-479f-be4a-4c9975a83268\") " Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.170855 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities" (OuterVolumeSpecName: "utilities") pod "a870983f-fa46-479f-be4a-4c9975a83268" (UID: "a870983f-fa46-479f-be4a-4c9975a83268"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.170954 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.175730 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z" (OuterVolumeSpecName: "kube-api-access-hsq7z") pod "a870983f-fa46-479f-be4a-4c9975a83268" (UID: "a870983f-fa46-479f-be4a-4c9975a83268"). InnerVolumeSpecName "kube-api-access-hsq7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.232625 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a870983f-fa46-479f-be4a-4c9975a83268" (UID: "a870983f-fa46-479f-be4a-4c9975a83268"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.274096 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a870983f-fa46-479f-be4a-4c9975a83268-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.274329 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsq7z\" (UniqueName: \"kubernetes.io/projected/a870983f-fa46-479f-be4a-4c9975a83268-kube-api-access-hsq7z\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.284539 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:45 crc kubenswrapper[4754]: W1011 03:40:45.294354 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01286d3c_ccb4_4245_9228_2e1a4ed0ebc0.slice/crio-f45cfce4f71982d722adf0952b0b94ceb275cb4bd0fff9c1260515f3bafeffad WatchSource:0}: Error finding container f45cfce4f71982d722adf0952b0b94ceb275cb4bd0fff9c1260515f3bafeffad: Status 404 returned error can't find the container with id f45cfce4f71982d722adf0952b0b94ceb275cb4bd0fff9c1260515f3bafeffad Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.332928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerStarted","Data":"f45cfce4f71982d722adf0952b0b94ceb275cb4bd0fff9c1260515f3bafeffad"} Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.340529 4754 generic.go:334] "Generic (PLEG): container finished" podID="a870983f-fa46-479f-be4a-4c9975a83268" containerID="753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777" exitCode=0 Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.340658 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerDied","Data":"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777"} Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.340743 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wr9zv" event={"ID":"a870983f-fa46-479f-be4a-4c9975a83268","Type":"ContainerDied","Data":"d02f64f920ee4173a871d8fdef6ef24f3e457544b76a7342ec015140bd894a3e"} Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.340817 4754 scope.go:117] "RemoveContainer" containerID="753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.341000 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wr9zv" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.379478 4754 scope.go:117] "RemoveContainer" containerID="a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.389272 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.405044 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wr9zv"] Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.416631 4754 scope.go:117] "RemoveContainer" containerID="2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.434438 4754 scope.go:117] "RemoveContainer" containerID="753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777" Oct 11 03:40:45 crc kubenswrapper[4754]: E1011 03:40:45.434866 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777\": container with ID starting with 753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777 not found: ID does not exist" containerID="753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.434898 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777"} err="failed to get container status \"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777\": rpc error: code = NotFound desc = could not find container \"753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777\": container with ID starting with 753c476b2a8b2d1e92d7eadfe21ee9a355349a28ad65f0458afe074644cfa777 not found: ID does not exist" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.434921 4754 scope.go:117] "RemoveContainer" containerID="a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c" Oct 11 03:40:45 crc kubenswrapper[4754]: E1011 03:40:45.435192 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c\": container with ID starting with a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c not found: ID does not exist" containerID="a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.435215 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c"} err="failed to get container status \"a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c\": rpc error: code = NotFound desc = could not find container \"a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c\": container with ID starting with a2dce70a054d6b143e0628be42fb1166b9a42b6042be08f93cecd01a1d262b3c not found: ID does not exist" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.435235 4754 scope.go:117] "RemoveContainer" containerID="2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14" Oct 11 03:40:45 crc kubenswrapper[4754]: E1011 03:40:45.435475 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14\": container with ID starting with 2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14 not found: ID does not exist" containerID="2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14" Oct 11 03:40:45 crc kubenswrapper[4754]: I1011 03:40:45.435497 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14"} err="failed to get container status \"2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14\": rpc error: code = NotFound desc = could not find container \"2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14\": container with ID starting with 2b645a5c9978f5edf84354716226e61247b10f846a5e80f9908bf6ddf521ff14 not found: ID does not exist" Oct 11 03:40:46 crc kubenswrapper[4754]: I1011 03:40:46.359074 4754 generic.go:334] "Generic (PLEG): container finished" podID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerID="334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362" exitCode=0 Oct 11 03:40:46 crc kubenswrapper[4754]: I1011 03:40:46.359173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerDied","Data":"334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362"} Oct 11 03:40:47 crc kubenswrapper[4754]: I1011 03:40:47.095285 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a870983f-fa46-479f-be4a-4c9975a83268" path="/var/lib/kubelet/pods/a870983f-fa46-479f-be4a-4c9975a83268/volumes" Oct 11 03:40:47 crc kubenswrapper[4754]: I1011 03:40:47.367669 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerStarted","Data":"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d"} Oct 11 03:40:48 crc kubenswrapper[4754]: I1011 03:40:48.379652 4754 generic.go:334] "Generic (PLEG): container finished" podID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerID="8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d" exitCode=0 Oct 11 03:40:48 crc kubenswrapper[4754]: I1011 03:40:48.379707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerDied","Data":"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d"} Oct 11 03:40:49 crc kubenswrapper[4754]: I1011 03:40:49.393038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerStarted","Data":"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700"} Oct 11 03:40:49 crc kubenswrapper[4754]: I1011 03:40:49.423330 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4sq6k" podStartSLOduration=2.896924487 podStartE2EDuration="5.423305752s" podCreationTimestamp="2025-10-11 03:40:44 +0000 UTC" firstStartedPulling="2025-10-11 03:40:46.361256897 +0000 UTC m=+2093.920201682" lastFinishedPulling="2025-10-11 03:40:48.887638162 +0000 UTC m=+2096.446582947" observedRunningTime="2025-10-11 03:40:49.416007606 +0000 UTC m=+2096.974952391" watchObservedRunningTime="2025-10-11 03:40:49.423305752 +0000 UTC m=+2096.982250547" Oct 11 03:40:54 crc kubenswrapper[4754]: I1011 03:40:54.784106 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:54 crc kubenswrapper[4754]: I1011 03:40:54.784481 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:54 crc kubenswrapper[4754]: I1011 03:40:54.830395 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:55 crc kubenswrapper[4754]: I1011 03:40:55.525061 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:55 crc kubenswrapper[4754]: I1011 03:40:55.570347 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:57 crc kubenswrapper[4754]: I1011 03:40:57.455175 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4sq6k" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="registry-server" containerID="cri-o://4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700" gracePeriod=2 Oct 11 03:40:57 crc kubenswrapper[4754]: I1011 03:40:57.901034 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.002183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content\") pod \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.002383 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4mjb\" (UniqueName: \"kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb\") pod \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.002439 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities\") pod \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\" (UID: \"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0\") " Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.003814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities" (OuterVolumeSpecName: "utilities") pod "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" (UID: "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.008548 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb" (OuterVolumeSpecName: "kube-api-access-p4mjb") pod "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" (UID: "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0"). InnerVolumeSpecName "kube-api-access-p4mjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.085361 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" (UID: "01286d3c-ccb4-4245-9228-2e1a4ed0ebc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.104353 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.104386 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4mjb\" (UniqueName: \"kubernetes.io/projected/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-kube-api-access-p4mjb\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.104398 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.468805 4754 generic.go:334] "Generic (PLEG): container finished" podID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerID="4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700" exitCode=0 Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.468854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerDied","Data":"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700"} Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.468882 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4sq6k" event={"ID":"01286d3c-ccb4-4245-9228-2e1a4ed0ebc0","Type":"ContainerDied","Data":"f45cfce4f71982d722adf0952b0b94ceb275cb4bd0fff9c1260515f3bafeffad"} Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.468901 4754 scope.go:117] "RemoveContainer" containerID="4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.468915 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4sq6k" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.516294 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.516776 4754 scope.go:117] "RemoveContainer" containerID="8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.522831 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4sq6k"] Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.536119 4754 scope.go:117] "RemoveContainer" containerID="334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.579867 4754 scope.go:117] "RemoveContainer" containerID="4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700" Oct 11 03:40:58 crc kubenswrapper[4754]: E1011 03:40:58.580475 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700\": container with ID starting with 4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700 not found: ID does not exist" containerID="4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.580530 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700"} err="failed to get container status \"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700\": rpc error: code = NotFound desc = could not find container \"4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700\": container with ID starting with 4563ea7abae01a9a8b4e7e43c7222cf4dfc283e52c6e99af655b474e1ad3f700 not found: ID does not exist" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.580573 4754 scope.go:117] "RemoveContainer" containerID="8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d" Oct 11 03:40:58 crc kubenswrapper[4754]: E1011 03:40:58.581016 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d\": container with ID starting with 8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d not found: ID does not exist" containerID="8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.581045 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d"} err="failed to get container status \"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d\": rpc error: code = NotFound desc = could not find container \"8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d\": container with ID starting with 8b67ccd946116a5e64c8c662b52194859b8ff3f4c375a31636fc64033fd67a3d not found: ID does not exist" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.581067 4754 scope.go:117] "RemoveContainer" containerID="334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362" Oct 11 03:40:58 crc kubenswrapper[4754]: E1011 03:40:58.581403 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362\": container with ID starting with 334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362 not found: ID does not exist" containerID="334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362" Oct 11 03:40:58 crc kubenswrapper[4754]: I1011 03:40:58.581429 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362"} err="failed to get container status \"334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362\": rpc error: code = NotFound desc = could not find container \"334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362\": container with ID starting with 334b9647feddeb69ed26cafa13f35d669b761af9b23cf82a6c5b93627aaaf362 not found: ID does not exist" Oct 11 03:40:59 crc kubenswrapper[4754]: I1011 03:40:59.106079 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" path="/var/lib/kubelet/pods/01286d3c-ccb4-4245-9228-2e1a4ed0ebc0/volumes" Oct 11 03:41:00 crc kubenswrapper[4754]: I1011 03:41:00.736252 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:41:00 crc kubenswrapper[4754]: I1011 03:41:00.736498 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:41:05 crc kubenswrapper[4754]: I1011 03:41:05.827245 4754 scope.go:117] "RemoveContainer" containerID="dd1600c607668d9198bdc0e6b7e4b1931902b073c8c10381a0e819436d295504" Oct 11 03:41:05 crc kubenswrapper[4754]: I1011 03:41:05.868604 4754 scope.go:117] "RemoveContainer" containerID="e4376dc3e7a4540db08e27e856b82300ccffc59453e78805b2264ce66a5cbd96" Oct 11 03:41:08 crc kubenswrapper[4754]: I1011 03:41:08.539305 4754 generic.go:334] "Generic (PLEG): container finished" podID="f77837ec-b2bb-4683-bc5f-21ec5145910c" containerID="fd3dd95860e9667b262c73f21c6b98864c7df335f650f7a6bf6f96f09745971b" exitCode=0 Oct 11 03:41:08 crc kubenswrapper[4754]: I1011 03:41:08.539411 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" event={"ID":"f77837ec-b2bb-4683-bc5f-21ec5145910c","Type":"ContainerDied","Data":"fd3dd95860e9667b262c73f21c6b98864c7df335f650f7a6bf6f96f09745971b"} Oct 11 03:41:09 crc kubenswrapper[4754]: I1011 03:41:09.973212 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.123494 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key\") pod \"f77837ec-b2bb-4683-bc5f-21ec5145910c\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.123630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory\") pod \"f77837ec-b2bb-4683-bc5f-21ec5145910c\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.123748 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs6f5\" (UniqueName: \"kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5\") pod \"f77837ec-b2bb-4683-bc5f-21ec5145910c\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.123780 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle\") pod \"f77837ec-b2bb-4683-bc5f-21ec5145910c\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.123824 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph\") pod \"f77837ec-b2bb-4683-bc5f-21ec5145910c\" (UID: \"f77837ec-b2bb-4683-bc5f-21ec5145910c\") " Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.133174 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f77837ec-b2bb-4683-bc5f-21ec5145910c" (UID: "f77837ec-b2bb-4683-bc5f-21ec5145910c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.133212 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5" (OuterVolumeSpecName: "kube-api-access-xs6f5") pod "f77837ec-b2bb-4683-bc5f-21ec5145910c" (UID: "f77837ec-b2bb-4683-bc5f-21ec5145910c"). InnerVolumeSpecName "kube-api-access-xs6f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.133207 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph" (OuterVolumeSpecName: "ceph") pod "f77837ec-b2bb-4683-bc5f-21ec5145910c" (UID: "f77837ec-b2bb-4683-bc5f-21ec5145910c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.150003 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory" (OuterVolumeSpecName: "inventory") pod "f77837ec-b2bb-4683-bc5f-21ec5145910c" (UID: "f77837ec-b2bb-4683-bc5f-21ec5145910c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.173806 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f77837ec-b2bb-4683-bc5f-21ec5145910c" (UID: "f77837ec-b2bb-4683-bc5f-21ec5145910c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.225590 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.225626 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.225637 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs6f5\" (UniqueName: \"kubernetes.io/projected/f77837ec-b2bb-4683-bc5f-21ec5145910c-kube-api-access-xs6f5\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.225650 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.225661 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f77837ec-b2bb-4683-bc5f-21ec5145910c-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.557193 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" event={"ID":"f77837ec-b2bb-4683-bc5f-21ec5145910c","Type":"ContainerDied","Data":"c22a2e67b22e5d2f0a6e50045e3855c4482d2485182c4d99717890b57dafad51"} Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.557229 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c22a2e67b22e5d2f0a6e50045e3855c4482d2485182c4d99717890b57dafad51" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.557273 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.646663 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68"] Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647057 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="extract-utilities" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647077 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="extract-utilities" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647107 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647116 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647148 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77837ec-b2bb-4683-bc5f-21ec5145910c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647158 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77837ec-b2bb-4683-bc5f-21ec5145910c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647173 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="extract-content" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647181 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="extract-content" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647192 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="extract-utilities" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647201 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="extract-utilities" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647215 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647222 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: E1011 03:41:10.647229 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="extract-content" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647237 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="extract-content" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647439 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a870983f-fa46-479f-be4a-4c9975a83268" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647460 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77837ec-b2bb-4683-bc5f-21ec5145910c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.647475 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="01286d3c-ccb4-4245-9228-2e1a4ed0ebc0" containerName="registry-server" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.648210 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.652420 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.652716 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.652889 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.653030 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.653157 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.683769 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68"] Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.733896 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.733944 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.734004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxlf8\" (UniqueName: \"kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.734072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.835818 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.835919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.835947 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.836002 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxlf8\" (UniqueName: \"kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.839569 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.840655 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.843499 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.851413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxlf8\" (UniqueName: \"kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bds68\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:10 crc kubenswrapper[4754]: I1011 03:41:10.964567 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:11 crc kubenswrapper[4754]: I1011 03:41:11.488030 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68"] Oct 11 03:41:11 crc kubenswrapper[4754]: I1011 03:41:11.567114 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" event={"ID":"dd2df864-0dde-4eaa-88f4-ae2e093087ec","Type":"ContainerStarted","Data":"b9697f95ec321b50b4ba10a6468ae29bd2686fc54994b5586d53cae69070c7bc"} Oct 11 03:41:12 crc kubenswrapper[4754]: I1011 03:41:12.581680 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" event={"ID":"dd2df864-0dde-4eaa-88f4-ae2e093087ec","Type":"ContainerStarted","Data":"ca29f982c0f9381127a3082014fd509e561273c34d34273f19df90fafd9327c5"} Oct 11 03:41:12 crc kubenswrapper[4754]: I1011 03:41:12.600721 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" podStartSLOduration=2.079356916 podStartE2EDuration="2.600701411s" podCreationTimestamp="2025-10-11 03:41:10 +0000 UTC" firstStartedPulling="2025-10-11 03:41:11.49468056 +0000 UTC m=+2119.053625345" lastFinishedPulling="2025-10-11 03:41:12.016025055 +0000 UTC m=+2119.574969840" observedRunningTime="2025-10-11 03:41:12.596467652 +0000 UTC m=+2120.155412437" watchObservedRunningTime="2025-10-11 03:41:12.600701411 +0000 UTC m=+2120.159646206" Oct 11 03:41:30 crc kubenswrapper[4754]: I1011 03:41:30.736158 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:41:30 crc kubenswrapper[4754]: I1011 03:41:30.737063 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.432101 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.437090 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.447126 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.577811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.578041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxk49\" (UniqueName: \"kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.578235 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.679467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.679515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxk49\" (UniqueName: \"kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.679602 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.680168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.680180 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.704294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxk49\" (UniqueName: \"kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49\") pod \"redhat-marketplace-rfrxg\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.775730 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.775901 4754 generic.go:334] "Generic (PLEG): container finished" podID="dd2df864-0dde-4eaa-88f4-ae2e093087ec" containerID="ca29f982c0f9381127a3082014fd509e561273c34d34273f19df90fafd9327c5" exitCode=0 Oct 11 03:41:37 crc kubenswrapper[4754]: I1011 03:41:37.775943 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" event={"ID":"dd2df864-0dde-4eaa-88f4-ae2e093087ec","Type":"ContainerDied","Data":"ca29f982c0f9381127a3082014fd509e561273c34d34273f19df90fafd9327c5"} Oct 11 03:41:38 crc kubenswrapper[4754]: I1011 03:41:38.203701 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:38 crc kubenswrapper[4754]: I1011 03:41:38.785410 4754 generic.go:334] "Generic (PLEG): container finished" podID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerID="752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974" exitCode=0 Oct 11 03:41:38 crc kubenswrapper[4754]: I1011 03:41:38.785518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerDied","Data":"752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974"} Oct 11 03:41:38 crc kubenswrapper[4754]: I1011 03:41:38.785755 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerStarted","Data":"f78a9a2819e9f3d1b69f622fdfbdb5a094dff44ae3af8a90a3c44579306dd125"} Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.160923 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.320066 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph\") pod \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.320174 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key\") pod \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.320198 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory\") pod \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.320307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxlf8\" (UniqueName: \"kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8\") pod \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\" (UID: \"dd2df864-0dde-4eaa-88f4-ae2e093087ec\") " Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.325598 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8" (OuterVolumeSpecName: "kube-api-access-gxlf8") pod "dd2df864-0dde-4eaa-88f4-ae2e093087ec" (UID: "dd2df864-0dde-4eaa-88f4-ae2e093087ec"). InnerVolumeSpecName "kube-api-access-gxlf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.325650 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph" (OuterVolumeSpecName: "ceph") pod "dd2df864-0dde-4eaa-88f4-ae2e093087ec" (UID: "dd2df864-0dde-4eaa-88f4-ae2e093087ec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.347353 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory" (OuterVolumeSpecName: "inventory") pod "dd2df864-0dde-4eaa-88f4-ae2e093087ec" (UID: "dd2df864-0dde-4eaa-88f4-ae2e093087ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.359781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd2df864-0dde-4eaa-88f4-ae2e093087ec" (UID: "dd2df864-0dde-4eaa-88f4-ae2e093087ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.422115 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.422151 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.422163 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd2df864-0dde-4eaa-88f4-ae2e093087ec-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.422174 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxlf8\" (UniqueName: \"kubernetes.io/projected/dd2df864-0dde-4eaa-88f4-ae2e093087ec-kube-api-access-gxlf8\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.793884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" event={"ID":"dd2df864-0dde-4eaa-88f4-ae2e093087ec","Type":"ContainerDied","Data":"b9697f95ec321b50b4ba10a6468ae29bd2686fc54994b5586d53cae69070c7bc"} Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.793929 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9697f95ec321b50b4ba10a6468ae29bd2686fc54994b5586d53cae69070c7bc" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.793906 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bds68" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.795587 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerStarted","Data":"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95"} Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.879880 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv"] Oct 11 03:41:39 crc kubenswrapper[4754]: E1011 03:41:39.880295 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2df864-0dde-4eaa-88f4-ae2e093087ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.880322 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2df864-0dde-4eaa-88f4-ae2e093087ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.880534 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2df864-0dde-4eaa-88f4-ae2e093087ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.881405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.883341 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.883418 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.883794 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.883865 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.884950 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.894331 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv"] Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.930100 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.930208 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.930260 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l949f\" (UniqueName: \"kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:39 crc kubenswrapper[4754]: I1011 03:41:39.930312 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.031565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.031919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l949f\" (UniqueName: \"kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.031982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.032039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.036893 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.036931 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.037496 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.047216 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l949f\" (UniqueName: \"kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4wllv\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.210027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.720436 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv"] Oct 11 03:41:40 crc kubenswrapper[4754]: W1011 03:41:40.722341 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e956480_ea42_4f8b_96ce_cebce380af8a.slice/crio-7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3 WatchSource:0}: Error finding container 7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3: Status 404 returned error can't find the container with id 7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3 Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.806769 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" event={"ID":"4e956480-ea42-4f8b-96ce-cebce380af8a","Type":"ContainerStarted","Data":"7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3"} Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.810528 4754 generic.go:334] "Generic (PLEG): container finished" podID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerID="9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95" exitCode=0 Oct 11 03:41:40 crc kubenswrapper[4754]: I1011 03:41:40.810566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerDied","Data":"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95"} Oct 11 03:41:41 crc kubenswrapper[4754]: I1011 03:41:41.822390 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerStarted","Data":"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f"} Oct 11 03:41:41 crc kubenswrapper[4754]: I1011 03:41:41.832764 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" event={"ID":"4e956480-ea42-4f8b-96ce-cebce380af8a","Type":"ContainerStarted","Data":"1e8d4bb8cefe83e78b90580f20440e24dcaedb232922b43b4b7259cb8879a38f"} Oct 11 03:41:41 crc kubenswrapper[4754]: I1011 03:41:41.857756 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rfrxg" podStartSLOduration=2.342088965 podStartE2EDuration="4.857735358s" podCreationTimestamp="2025-10-11 03:41:37 +0000 UTC" firstStartedPulling="2025-10-11 03:41:38.787119486 +0000 UTC m=+2146.346064261" lastFinishedPulling="2025-10-11 03:41:41.302765869 +0000 UTC m=+2148.861710654" observedRunningTime="2025-10-11 03:41:41.852560784 +0000 UTC m=+2149.411505589" watchObservedRunningTime="2025-10-11 03:41:41.857735358 +0000 UTC m=+2149.416680153" Oct 11 03:41:41 crc kubenswrapper[4754]: I1011 03:41:41.876877 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" podStartSLOduration=2.413496181 podStartE2EDuration="2.876855678s" podCreationTimestamp="2025-10-11 03:41:39 +0000 UTC" firstStartedPulling="2025-10-11 03:41:40.72537255 +0000 UTC m=+2148.284317355" lastFinishedPulling="2025-10-11 03:41:41.188732067 +0000 UTC m=+2148.747676852" observedRunningTime="2025-10-11 03:41:41.871807108 +0000 UTC m=+2149.430751903" watchObservedRunningTime="2025-10-11 03:41:41.876855678 +0000 UTC m=+2149.435800463" Oct 11 03:41:46 crc kubenswrapper[4754]: E1011 03:41:46.483449 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e956480_ea42_4f8b_96ce_cebce380af8a.slice/crio-conmon-1e8d4bb8cefe83e78b90580f20440e24dcaedb232922b43b4b7259cb8879a38f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e956480_ea42_4f8b_96ce_cebce380af8a.slice/crio-1e8d4bb8cefe83e78b90580f20440e24dcaedb232922b43b4b7259cb8879a38f.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:41:46 crc kubenswrapper[4754]: I1011 03:41:46.877917 4754 generic.go:334] "Generic (PLEG): container finished" podID="4e956480-ea42-4f8b-96ce-cebce380af8a" containerID="1e8d4bb8cefe83e78b90580f20440e24dcaedb232922b43b4b7259cb8879a38f" exitCode=0 Oct 11 03:41:46 crc kubenswrapper[4754]: I1011 03:41:46.878010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" event={"ID":"4e956480-ea42-4f8b-96ce-cebce380af8a","Type":"ContainerDied","Data":"1e8d4bb8cefe83e78b90580f20440e24dcaedb232922b43b4b7259cb8879a38f"} Oct 11 03:41:47 crc kubenswrapper[4754]: I1011 03:41:47.775948 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:47 crc kubenswrapper[4754]: I1011 03:41:47.776071 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:47 crc kubenswrapper[4754]: I1011 03:41:47.822467 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:47 crc kubenswrapper[4754]: I1011 03:41:47.938214 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.323524 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.480342 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph\") pod \"4e956480-ea42-4f8b-96ce-cebce380af8a\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.481027 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l949f\" (UniqueName: \"kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f\") pod \"4e956480-ea42-4f8b-96ce-cebce380af8a\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.481168 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory\") pod \"4e956480-ea42-4f8b-96ce-cebce380af8a\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.481227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key\") pod \"4e956480-ea42-4f8b-96ce-cebce380af8a\" (UID: \"4e956480-ea42-4f8b-96ce-cebce380af8a\") " Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.485996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph" (OuterVolumeSpecName: "ceph") pod "4e956480-ea42-4f8b-96ce-cebce380af8a" (UID: "4e956480-ea42-4f8b-96ce-cebce380af8a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.486430 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f" (OuterVolumeSpecName: "kube-api-access-l949f") pod "4e956480-ea42-4f8b-96ce-cebce380af8a" (UID: "4e956480-ea42-4f8b-96ce-cebce380af8a"). InnerVolumeSpecName "kube-api-access-l949f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.506085 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e956480-ea42-4f8b-96ce-cebce380af8a" (UID: "4e956480-ea42-4f8b-96ce-cebce380af8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.528089 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory" (OuterVolumeSpecName: "inventory") pod "4e956480-ea42-4f8b-96ce-cebce380af8a" (UID: "4e956480-ea42-4f8b-96ce-cebce380af8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.583044 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l949f\" (UniqueName: \"kubernetes.io/projected/4e956480-ea42-4f8b-96ce-cebce380af8a-kube-api-access-l949f\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.583082 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.583095 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.583107 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e956480-ea42-4f8b-96ce-cebce380af8a-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.899109 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.899118 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4wllv" event={"ID":"4e956480-ea42-4f8b-96ce-cebce380af8a","Type":"ContainerDied","Data":"7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3"} Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.900172 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c04d1244f5385cfbe31b1a3e53584b3d6a0484164c87282d21c762f49c2fca3" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.973774 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj"] Oct 11 03:41:48 crc kubenswrapper[4754]: E1011 03:41:48.974154 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e956480-ea42-4f8b-96ce-cebce380af8a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.974172 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e956480-ea42-4f8b-96ce-cebce380af8a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.974343 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e956480-ea42-4f8b-96ce-cebce380af8a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.974918 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.977179 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.977227 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.977897 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.978561 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.979661 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:41:48 crc kubenswrapper[4754]: I1011 03:41:48.994565 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj"] Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.093887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.093954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.094619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2287z\" (UniqueName: \"kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.095249 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.196922 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.197004 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.197056 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2287z\" (UniqueName: \"kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.197236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.203536 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.209029 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.209627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.211146 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.214877 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2287z\" (UniqueName: \"kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hjzrj\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.291861 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.790143 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj"] Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.909253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" event={"ID":"9a11ae05-4ae6-48a9-980d-84a802d36fb5","Type":"ContainerStarted","Data":"1dcd3d442a1e97c7981275cd9365024f1ff6521eef80aa5aea373ecdb5a0ebac"} Oct 11 03:41:49 crc kubenswrapper[4754]: I1011 03:41:49.909351 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rfrxg" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="registry-server" containerID="cri-o://29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f" gracePeriod=2 Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.361458 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.519893 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities\") pod \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.519958 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxk49\" (UniqueName: \"kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49\") pod \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.520037 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content\") pod \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\" (UID: \"021d3ce5-d907-426b-9dd5-8f4f2f5f709e\") " Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.520918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities" (OuterVolumeSpecName: "utilities") pod "021d3ce5-d907-426b-9dd5-8f4f2f5f709e" (UID: "021d3ce5-d907-426b-9dd5-8f4f2f5f709e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.525558 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49" (OuterVolumeSpecName: "kube-api-access-vxk49") pod "021d3ce5-d907-426b-9dd5-8f4f2f5f709e" (UID: "021d3ce5-d907-426b-9dd5-8f4f2f5f709e"). InnerVolumeSpecName "kube-api-access-vxk49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.534317 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "021d3ce5-d907-426b-9dd5-8f4f2f5f709e" (UID: "021d3ce5-d907-426b-9dd5-8f4f2f5f709e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.621777 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.621819 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxk49\" (UniqueName: \"kubernetes.io/projected/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-kube-api-access-vxk49\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.621830 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/021d3ce5-d907-426b-9dd5-8f4f2f5f709e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.919880 4754 generic.go:334] "Generic (PLEG): container finished" podID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerID="29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f" exitCode=0 Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.919933 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfrxg" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.919956 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerDied","Data":"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f"} Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.920000 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfrxg" event={"ID":"021d3ce5-d907-426b-9dd5-8f4f2f5f709e","Type":"ContainerDied","Data":"f78a9a2819e9f3d1b69f622fdfbdb5a094dff44ae3af8a90a3c44579306dd125"} Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.920025 4754 scope.go:117] "RemoveContainer" containerID="29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.923824 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" event={"ID":"9a11ae05-4ae6-48a9-980d-84a802d36fb5","Type":"ContainerStarted","Data":"ba9746ccc689bfd6f15a5be1ff4c884a762365a4de493b1bce7d4a85c7763720"} Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.945390 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" podStartSLOduration=2.519702433 podStartE2EDuration="2.945373166s" podCreationTimestamp="2025-10-11 03:41:48 +0000 UTC" firstStartedPulling="2025-10-11 03:41:49.795550493 +0000 UTC m=+2157.354495278" lastFinishedPulling="2025-10-11 03:41:50.221221226 +0000 UTC m=+2157.780166011" observedRunningTime="2025-10-11 03:41:50.940267964 +0000 UTC m=+2158.499212749" watchObservedRunningTime="2025-10-11 03:41:50.945373166 +0000 UTC m=+2158.504317951" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.955657 4754 scope.go:117] "RemoveContainer" containerID="9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95" Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.961489 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.976856 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfrxg"] Oct 11 03:41:50 crc kubenswrapper[4754]: I1011 03:41:50.977077 4754 scope.go:117] "RemoveContainer" containerID="752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.014595 4754 scope.go:117] "RemoveContainer" containerID="29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f" Oct 11 03:41:51 crc kubenswrapper[4754]: E1011 03:41:51.015133 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f\": container with ID starting with 29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f not found: ID does not exist" containerID="29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.015176 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f"} err="failed to get container status \"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f\": rpc error: code = NotFound desc = could not find container \"29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f\": container with ID starting with 29fa052d4a17d54d56e3c5e7253fd2902d7c238fd86d29b0bed0cf7e25c8614f not found: ID does not exist" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.015202 4754 scope.go:117] "RemoveContainer" containerID="9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95" Oct 11 03:41:51 crc kubenswrapper[4754]: E1011 03:41:51.015495 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95\": container with ID starting with 9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95 not found: ID does not exist" containerID="9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.015518 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95"} err="failed to get container status \"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95\": rpc error: code = NotFound desc = could not find container \"9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95\": container with ID starting with 9f7185d7d089a53e036ac1a5078206286dec305fe630e24900d295874ae14d95 not found: ID does not exist" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.015531 4754 scope.go:117] "RemoveContainer" containerID="752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974" Oct 11 03:41:51 crc kubenswrapper[4754]: E1011 03:41:51.015750 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974\": container with ID starting with 752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974 not found: ID does not exist" containerID="752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.015767 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974"} err="failed to get container status \"752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974\": rpc error: code = NotFound desc = could not find container \"752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974\": container with ID starting with 752bc8d827262725caa804a33cd41fb8c57af66f0e98146f9b726fcac689e974 not found: ID does not exist" Oct 11 03:41:51 crc kubenswrapper[4754]: I1011 03:41:51.095561 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" path="/var/lib/kubelet/pods/021d3ce5-d907-426b-9dd5-8f4f2f5f709e/volumes" Oct 11 03:42:00 crc kubenswrapper[4754]: I1011 03:42:00.736620 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:42:00 crc kubenswrapper[4754]: I1011 03:42:00.737189 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:42:00 crc kubenswrapper[4754]: I1011 03:42:00.737242 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:42:00 crc kubenswrapper[4754]: I1011 03:42:00.739390 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:42:00 crc kubenswrapper[4754]: I1011 03:42:00.739822 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a" gracePeriod=600 Oct 11 03:42:01 crc kubenswrapper[4754]: I1011 03:42:01.009293 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a" exitCode=0 Oct 11 03:42:01 crc kubenswrapper[4754]: I1011 03:42:01.009338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a"} Oct 11 03:42:01 crc kubenswrapper[4754]: I1011 03:42:01.009371 4754 scope.go:117] "RemoveContainer" containerID="28e421f12ab6803ceae3cbdb6d82851e55813aeec70a6b08c60746bb38fee1c5" Oct 11 03:42:02 crc kubenswrapper[4754]: I1011 03:42:02.019199 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6"} Oct 11 03:42:28 crc kubenswrapper[4754]: I1011 03:42:28.221626 4754 generic.go:334] "Generic (PLEG): container finished" podID="9a11ae05-4ae6-48a9-980d-84a802d36fb5" containerID="ba9746ccc689bfd6f15a5be1ff4c884a762365a4de493b1bce7d4a85c7763720" exitCode=0 Oct 11 03:42:28 crc kubenswrapper[4754]: I1011 03:42:28.221800 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" event={"ID":"9a11ae05-4ae6-48a9-980d-84a802d36fb5","Type":"ContainerDied","Data":"ba9746ccc689bfd6f15a5be1ff4c884a762365a4de493b1bce7d4a85c7763720"} Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.610781 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.740484 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key\") pod \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.740819 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph\") pod \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.740862 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2287z\" (UniqueName: \"kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z\") pod \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.740933 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory\") pod \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\" (UID: \"9a11ae05-4ae6-48a9-980d-84a802d36fb5\") " Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.748104 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z" (OuterVolumeSpecName: "kube-api-access-2287z") pod "9a11ae05-4ae6-48a9-980d-84a802d36fb5" (UID: "9a11ae05-4ae6-48a9-980d-84a802d36fb5"). InnerVolumeSpecName "kube-api-access-2287z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.761142 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph" (OuterVolumeSpecName: "ceph") pod "9a11ae05-4ae6-48a9-980d-84a802d36fb5" (UID: "9a11ae05-4ae6-48a9-980d-84a802d36fb5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.779644 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory" (OuterVolumeSpecName: "inventory") pod "9a11ae05-4ae6-48a9-980d-84a802d36fb5" (UID: "9a11ae05-4ae6-48a9-980d-84a802d36fb5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.793132 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a11ae05-4ae6-48a9-980d-84a802d36fb5" (UID: "9a11ae05-4ae6-48a9-980d-84a802d36fb5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.844726 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.845202 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.845250 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2287z\" (UniqueName: \"kubernetes.io/projected/9a11ae05-4ae6-48a9-980d-84a802d36fb5-kube-api-access-2287z\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:29 crc kubenswrapper[4754]: I1011 03:42:29.845263 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a11ae05-4ae6-48a9-980d-84a802d36fb5-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.239699 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" event={"ID":"9a11ae05-4ae6-48a9-980d-84a802d36fb5","Type":"ContainerDied","Data":"1dcd3d442a1e97c7981275cd9365024f1ff6521eef80aa5aea373ecdb5a0ebac"} Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.240159 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dcd3d442a1e97c7981275cd9365024f1ff6521eef80aa5aea373ecdb5a0ebac" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.240013 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hjzrj" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.315401 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv"] Oct 11 03:42:30 crc kubenswrapper[4754]: E1011 03:42:30.315713 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="registry-server" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.315727 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="registry-server" Oct 11 03:42:30 crc kubenswrapper[4754]: E1011 03:42:30.315740 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="extract-content" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.315746 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="extract-content" Oct 11 03:42:30 crc kubenswrapper[4754]: E1011 03:42:30.315773 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="extract-utilities" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.315779 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="extract-utilities" Oct 11 03:42:30 crc kubenswrapper[4754]: E1011 03:42:30.315795 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a11ae05-4ae6-48a9-980d-84a802d36fb5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.315803 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a11ae05-4ae6-48a9-980d-84a802d36fb5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.316013 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="021d3ce5-d907-426b-9dd5-8f4f2f5f709e" containerName="registry-server" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.316026 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a11ae05-4ae6-48a9-980d-84a802d36fb5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.316541 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.321568 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.321619 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.321585 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.321831 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.321895 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.331851 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv"] Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.457078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hh48\" (UniqueName: \"kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.457121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.457195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.457255 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.559130 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.559410 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hh48\" (UniqueName: \"kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.559492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.559625 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.563903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.564687 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.566993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.575693 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hh48\" (UniqueName: \"kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:30 crc kubenswrapper[4754]: I1011 03:42:30.635213 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:31 crc kubenswrapper[4754]: I1011 03:42:31.171150 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv"] Oct 11 03:42:31 crc kubenswrapper[4754]: I1011 03:42:31.248774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" event={"ID":"72c8febd-5b78-489c-9fdd-3235935e9934","Type":"ContainerStarted","Data":"a77c400008279ff874a4c1e64073a098b8222c6ed6f08ccef7472fe85648c245"} Oct 11 03:42:32 crc kubenswrapper[4754]: I1011 03:42:32.261454 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" event={"ID":"72c8febd-5b78-489c-9fdd-3235935e9934","Type":"ContainerStarted","Data":"61841a49a131f389741799b027ac97aff3ad6131ef9c120c43b8ec876aa7faa8"} Oct 11 03:42:32 crc kubenswrapper[4754]: I1011 03:42:32.299817 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" podStartSLOduration=1.7527453259999999 podStartE2EDuration="2.299778654s" podCreationTimestamp="2025-10-11 03:42:30 +0000 UTC" firstStartedPulling="2025-10-11 03:42:31.179538312 +0000 UTC m=+2198.738483087" lastFinishedPulling="2025-10-11 03:42:31.72657163 +0000 UTC m=+2199.285516415" observedRunningTime="2025-10-11 03:42:32.289087117 +0000 UTC m=+2199.848031932" watchObservedRunningTime="2025-10-11 03:42:32.299778654 +0000 UTC m=+2199.858723509" Oct 11 03:42:36 crc kubenswrapper[4754]: I1011 03:42:36.307700 4754 generic.go:334] "Generic (PLEG): container finished" podID="72c8febd-5b78-489c-9fdd-3235935e9934" containerID="61841a49a131f389741799b027ac97aff3ad6131ef9c120c43b8ec876aa7faa8" exitCode=0 Oct 11 03:42:36 crc kubenswrapper[4754]: I1011 03:42:36.307764 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" event={"ID":"72c8febd-5b78-489c-9fdd-3235935e9934","Type":"ContainerDied","Data":"61841a49a131f389741799b027ac97aff3ad6131ef9c120c43b8ec876aa7faa8"} Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.745642 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.920528 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hh48\" (UniqueName: \"kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48\") pod \"72c8febd-5b78-489c-9fdd-3235935e9934\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.920799 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory\") pod \"72c8febd-5b78-489c-9fdd-3235935e9934\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.920821 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph\") pod \"72c8febd-5b78-489c-9fdd-3235935e9934\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.920852 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key\") pod \"72c8febd-5b78-489c-9fdd-3235935e9934\" (UID: \"72c8febd-5b78-489c-9fdd-3235935e9934\") " Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.926285 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph" (OuterVolumeSpecName: "ceph") pod "72c8febd-5b78-489c-9fdd-3235935e9934" (UID: "72c8febd-5b78-489c-9fdd-3235935e9934"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.926324 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48" (OuterVolumeSpecName: "kube-api-access-2hh48") pod "72c8febd-5b78-489c-9fdd-3235935e9934" (UID: "72c8febd-5b78-489c-9fdd-3235935e9934"). InnerVolumeSpecName "kube-api-access-2hh48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.948134 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72c8febd-5b78-489c-9fdd-3235935e9934" (UID: "72c8febd-5b78-489c-9fdd-3235935e9934"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:37 crc kubenswrapper[4754]: I1011 03:42:37.952060 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory" (OuterVolumeSpecName: "inventory") pod "72c8febd-5b78-489c-9fdd-3235935e9934" (UID: "72c8febd-5b78-489c-9fdd-3235935e9934"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.023052 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hh48\" (UniqueName: \"kubernetes.io/projected/72c8febd-5b78-489c-9fdd-3235935e9934-kube-api-access-2hh48\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.023099 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.023118 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.023134 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72c8febd-5b78-489c-9fdd-3235935e9934-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.325891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" event={"ID":"72c8febd-5b78-489c-9fdd-3235935e9934","Type":"ContainerDied","Data":"a77c400008279ff874a4c1e64073a098b8222c6ed6f08ccef7472fe85648c245"} Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.326177 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a77c400008279ff874a4c1e64073a098b8222c6ed6f08ccef7472fe85648c245" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.325982 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.451710 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps"] Oct 11 03:42:38 crc kubenswrapper[4754]: E1011 03:42:38.452110 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c8febd-5b78-489c-9fdd-3235935e9934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.452127 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c8febd-5b78-489c-9fdd-3235935e9934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.452304 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c8febd-5b78-489c-9fdd-3235935e9934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.452883 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.455028 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.455237 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.455395 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.458831 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.466599 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.467074 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps"] Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.634311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.634382 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bld99\" (UniqueName: \"kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.634443 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.634519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.736712 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.736842 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.736881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bld99\" (UniqueName: \"kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.736903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.741078 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.741089 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.741428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.752350 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bld99\" (UniqueName: \"kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:38 crc kubenswrapper[4754]: I1011 03:42:38.766875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:42:39 crc kubenswrapper[4754]: I1011 03:42:39.250451 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps"] Oct 11 03:42:39 crc kubenswrapper[4754]: I1011 03:42:39.333979 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" event={"ID":"1d13d97f-00e5-4264-8817-6ae83b60819c","Type":"ContainerStarted","Data":"635566d43602705df805f76c647b28dead273c93261a8dde3f9f1ad3ff1d438b"} Oct 11 03:42:40 crc kubenswrapper[4754]: I1011 03:42:40.341396 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" event={"ID":"1d13d97f-00e5-4264-8817-6ae83b60819c","Type":"ContainerStarted","Data":"2aba756fe03b2ae2837c2cf5af64032896c95c38a9405990d7813800449ebb13"} Oct 11 03:42:40 crc kubenswrapper[4754]: I1011 03:42:40.365184 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" podStartSLOduration=1.933981344 podStartE2EDuration="2.3651709s" podCreationTimestamp="2025-10-11 03:42:38 +0000 UTC" firstStartedPulling="2025-10-11 03:42:39.253240848 +0000 UTC m=+2206.812185633" lastFinishedPulling="2025-10-11 03:42:39.684430404 +0000 UTC m=+2207.243375189" observedRunningTime="2025-10-11 03:42:40.361672153 +0000 UTC m=+2207.920616948" watchObservedRunningTime="2025-10-11 03:42:40.3651709 +0000 UTC m=+2207.924115685" Oct 11 03:43:28 crc kubenswrapper[4754]: I1011 03:43:28.723841 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d13d97f-00e5-4264-8817-6ae83b60819c" containerID="2aba756fe03b2ae2837c2cf5af64032896c95c38a9405990d7813800449ebb13" exitCode=0 Oct 11 03:43:28 crc kubenswrapper[4754]: I1011 03:43:28.723919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" event={"ID":"1d13d97f-00e5-4264-8817-6ae83b60819c","Type":"ContainerDied","Data":"2aba756fe03b2ae2837c2cf5af64032896c95c38a9405990d7813800449ebb13"} Oct 11 03:43:28 crc kubenswrapper[4754]: E1011 03:43:28.771782 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d13d97f_00e5_4264_8817_6ae83b60819c.slice/crio-conmon-2aba756fe03b2ae2837c2cf5af64032896c95c38a9405990d7813800449ebb13.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.120935 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.292307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph\") pod \"1d13d97f-00e5-4264-8817-6ae83b60819c\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.292655 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory\") pod \"1d13d97f-00e5-4264-8817-6ae83b60819c\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.292809 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bld99\" (UniqueName: \"kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99\") pod \"1d13d97f-00e5-4264-8817-6ae83b60819c\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.292891 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key\") pod \"1d13d97f-00e5-4264-8817-6ae83b60819c\" (UID: \"1d13d97f-00e5-4264-8817-6ae83b60819c\") " Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.298059 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph" (OuterVolumeSpecName: "ceph") pod "1d13d97f-00e5-4264-8817-6ae83b60819c" (UID: "1d13d97f-00e5-4264-8817-6ae83b60819c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.302163 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99" (OuterVolumeSpecName: "kube-api-access-bld99") pod "1d13d97f-00e5-4264-8817-6ae83b60819c" (UID: "1d13d97f-00e5-4264-8817-6ae83b60819c"). InnerVolumeSpecName "kube-api-access-bld99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.318382 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d13d97f-00e5-4264-8817-6ae83b60819c" (UID: "1d13d97f-00e5-4264-8817-6ae83b60819c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.324698 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory" (OuterVolumeSpecName: "inventory") pod "1d13d97f-00e5-4264-8817-6ae83b60819c" (UID: "1d13d97f-00e5-4264-8817-6ae83b60819c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.394530 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bld99\" (UniqueName: \"kubernetes.io/projected/1d13d97f-00e5-4264-8817-6ae83b60819c-kube-api-access-bld99\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.394570 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.394583 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.394593 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d13d97f-00e5-4264-8817-6ae83b60819c-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.743395 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" event={"ID":"1d13d97f-00e5-4264-8817-6ae83b60819c","Type":"ContainerDied","Data":"635566d43602705df805f76c647b28dead273c93261a8dde3f9f1ad3ff1d438b"} Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.743436 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="635566d43602705df805f76c647b28dead273c93261a8dde3f9f1ad3ff1d438b" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.743471 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.830355 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-89n62"] Oct 11 03:43:30 crc kubenswrapper[4754]: E1011 03:43:30.830755 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13d97f-00e5-4264-8817-6ae83b60819c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.830775 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13d97f-00e5-4264-8817-6ae83b60819c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.830936 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13d97f-00e5-4264-8817-6ae83b60819c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.831568 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.834501 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.834707 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.835017 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.835147 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.835693 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:43:30 crc kubenswrapper[4754]: I1011 03:43:30.836719 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-89n62"] Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.003888 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs2vz\" (UniqueName: \"kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.003947 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.004015 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.004070 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.105543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.105648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.105779 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs2vz\" (UniqueName: \"kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.105821 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.109200 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.110314 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.110663 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.120985 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs2vz\" (UniqueName: \"kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz\") pod \"ssh-known-hosts-edpm-deployment-89n62\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.186158 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.677480 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-89n62"] Oct 11 03:43:31 crc kubenswrapper[4754]: W1011 03:43:31.685232 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod511691f1_e3fa_4f9f_9a58_8a095ef1d39c.slice/crio-8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca WatchSource:0}: Error finding container 8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca: Status 404 returned error can't find the container with id 8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca Oct 11 03:43:31 crc kubenswrapper[4754]: I1011 03:43:31.752953 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" event={"ID":"511691f1-e3fa-4f9f-9a58-8a095ef1d39c","Type":"ContainerStarted","Data":"8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca"} Oct 11 03:43:32 crc kubenswrapper[4754]: I1011 03:43:32.761784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" event={"ID":"511691f1-e3fa-4f9f-9a58-8a095ef1d39c","Type":"ContainerStarted","Data":"a4c4641cb902e74841737f952d1f18af25b46c43ba57d5183bf0d10880ce7c2a"} Oct 11 03:43:32 crc kubenswrapper[4754]: I1011 03:43:32.782919 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" podStartSLOduration=2.197210107 podStartE2EDuration="2.782897316s" podCreationTimestamp="2025-10-11 03:43:30 +0000 UTC" firstStartedPulling="2025-10-11 03:43:31.68987837 +0000 UTC m=+2259.248823155" lastFinishedPulling="2025-10-11 03:43:32.275565559 +0000 UTC m=+2259.834510364" observedRunningTime="2025-10-11 03:43:32.778014601 +0000 UTC m=+2260.336959386" watchObservedRunningTime="2025-10-11 03:43:32.782897316 +0000 UTC m=+2260.341842101" Oct 11 03:43:42 crc kubenswrapper[4754]: I1011 03:43:42.833541 4754 generic.go:334] "Generic (PLEG): container finished" podID="511691f1-e3fa-4f9f-9a58-8a095ef1d39c" containerID="a4c4641cb902e74841737f952d1f18af25b46c43ba57d5183bf0d10880ce7c2a" exitCode=0 Oct 11 03:43:42 crc kubenswrapper[4754]: I1011 03:43:42.833623 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" event={"ID":"511691f1-e3fa-4f9f-9a58-8a095ef1d39c","Type":"ContainerDied","Data":"a4c4641cb902e74841737f952d1f18af25b46c43ba57d5183bf0d10880ce7c2a"} Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.213063 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.227026 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs2vz\" (UniqueName: \"kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz\") pod \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.227797 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam\") pod \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.227957 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph\") pod \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.228019 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0\") pod \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\" (UID: \"511691f1-e3fa-4f9f-9a58-8a095ef1d39c\") " Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.270794 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz" (OuterVolumeSpecName: "kube-api-access-hs2vz") pod "511691f1-e3fa-4f9f-9a58-8a095ef1d39c" (UID: "511691f1-e3fa-4f9f-9a58-8a095ef1d39c"). InnerVolumeSpecName "kube-api-access-hs2vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.273301 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "511691f1-e3fa-4f9f-9a58-8a095ef1d39c" (UID: "511691f1-e3fa-4f9f-9a58-8a095ef1d39c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.276169 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph" (OuterVolumeSpecName: "ceph") pod "511691f1-e3fa-4f9f-9a58-8a095ef1d39c" (UID: "511691f1-e3fa-4f9f-9a58-8a095ef1d39c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.277120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "511691f1-e3fa-4f9f-9a58-8a095ef1d39c" (UID: "511691f1-e3fa-4f9f-9a58-8a095ef1d39c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.330768 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs2vz\" (UniqueName: \"kubernetes.io/projected/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-kube-api-access-hs2vz\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.330805 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.330816 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.330833 4754 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/511691f1-e3fa-4f9f-9a58-8a095ef1d39c-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.850360 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" event={"ID":"511691f1-e3fa-4f9f-9a58-8a095ef1d39c","Type":"ContainerDied","Data":"8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca"} Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.850407 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8201d2835d23bbf675a56936ae9c4c260ae7312c3d80da4d0ed1d3a4b9ee9cca" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.850495 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-89n62" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.938635 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8"] Oct 11 03:43:44 crc kubenswrapper[4754]: E1011 03:43:44.939140 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511691f1-e3fa-4f9f-9a58-8a095ef1d39c" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.939157 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="511691f1-e3fa-4f9f-9a58-8a095ef1d39c" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.939401 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="511691f1-e3fa-4f9f-9a58-8a095ef1d39c" containerName="ssh-known-hosts-edpm-deployment" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.940547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.943227 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.943702 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.944115 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.944375 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.946661 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:43:44 crc kubenswrapper[4754]: I1011 03:43:44.953107 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8"] Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.043595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmj9b\" (UniqueName: \"kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.043636 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.043664 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.043710 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.145511 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmj9b\" (UniqueName: \"kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.145564 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.145587 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.146411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.150373 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.150464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.150879 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.162491 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmj9b\" (UniqueName: \"kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rsgf8\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.258808 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.795239 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8"] Oct 11 03:43:45 crc kubenswrapper[4754]: I1011 03:43:45.947650 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" event={"ID":"e0749e96-0b16-40c1-9f1c-18227b93168b","Type":"ContainerStarted","Data":"daf7f6b10cd55df3bf3b3140c3cbf3e217a50e2758b158871eec24b8f1bb983d"} Oct 11 03:43:46 crc kubenswrapper[4754]: I1011 03:43:46.958980 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" event={"ID":"e0749e96-0b16-40c1-9f1c-18227b93168b","Type":"ContainerStarted","Data":"cb6281020f28a11f03599efa34f99d900faefa971b9557aac673e181e3fdd59b"} Oct 11 03:43:46 crc kubenswrapper[4754]: I1011 03:43:46.982577 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" podStartSLOduration=2.271852634 podStartE2EDuration="2.98255928s" podCreationTimestamp="2025-10-11 03:43:44 +0000 UTC" firstStartedPulling="2025-10-11 03:43:45.804619299 +0000 UTC m=+2273.363564104" lastFinishedPulling="2025-10-11 03:43:46.515325945 +0000 UTC m=+2274.074270750" observedRunningTime="2025-10-11 03:43:46.978361614 +0000 UTC m=+2274.537306449" watchObservedRunningTime="2025-10-11 03:43:46.98255928 +0000 UTC m=+2274.541504075" Oct 11 03:43:55 crc kubenswrapper[4754]: I1011 03:43:55.029284 4754 generic.go:334] "Generic (PLEG): container finished" podID="e0749e96-0b16-40c1-9f1c-18227b93168b" containerID="cb6281020f28a11f03599efa34f99d900faefa971b9557aac673e181e3fdd59b" exitCode=0 Oct 11 03:43:55 crc kubenswrapper[4754]: I1011 03:43:55.029379 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" event={"ID":"e0749e96-0b16-40c1-9f1c-18227b93168b","Type":"ContainerDied","Data":"cb6281020f28a11f03599efa34f99d900faefa971b9557aac673e181e3fdd59b"} Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.419095 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.605840 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory\") pod \"e0749e96-0b16-40c1-9f1c-18227b93168b\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.606038 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmj9b\" (UniqueName: \"kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b\") pod \"e0749e96-0b16-40c1-9f1c-18227b93168b\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.606177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key\") pod \"e0749e96-0b16-40c1-9f1c-18227b93168b\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.606224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph\") pod \"e0749e96-0b16-40c1-9f1c-18227b93168b\" (UID: \"e0749e96-0b16-40c1-9f1c-18227b93168b\") " Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.611152 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph" (OuterVolumeSpecName: "ceph") pod "e0749e96-0b16-40c1-9f1c-18227b93168b" (UID: "e0749e96-0b16-40c1-9f1c-18227b93168b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.613325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b" (OuterVolumeSpecName: "kube-api-access-fmj9b") pod "e0749e96-0b16-40c1-9f1c-18227b93168b" (UID: "e0749e96-0b16-40c1-9f1c-18227b93168b"). InnerVolumeSpecName "kube-api-access-fmj9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.629880 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory" (OuterVolumeSpecName: "inventory") pod "e0749e96-0b16-40c1-9f1c-18227b93168b" (UID: "e0749e96-0b16-40c1-9f1c-18227b93168b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.637801 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e0749e96-0b16-40c1-9f1c-18227b93168b" (UID: "e0749e96-0b16-40c1-9f1c-18227b93168b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.708545 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmj9b\" (UniqueName: \"kubernetes.io/projected/e0749e96-0b16-40c1-9f1c-18227b93168b-kube-api-access-fmj9b\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.708854 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.709105 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:56 crc kubenswrapper[4754]: I1011 03:43:56.709264 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0749e96-0b16-40c1-9f1c-18227b93168b-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.047421 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" event={"ID":"e0749e96-0b16-40c1-9f1c-18227b93168b","Type":"ContainerDied","Data":"daf7f6b10cd55df3bf3b3140c3cbf3e217a50e2758b158871eec24b8f1bb983d"} Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.047458 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daf7f6b10cd55df3bf3b3140c3cbf3e217a50e2758b158871eec24b8f1bb983d" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.047726 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rsgf8" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.115948 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg"] Oct 11 03:43:57 crc kubenswrapper[4754]: E1011 03:43:57.116350 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0749e96-0b16-40c1-9f1c-18227b93168b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.116363 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0749e96-0b16-40c1-9f1c-18227b93168b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.116551 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0749e96-0b16-40c1-9f1c-18227b93168b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.117138 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.161847 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.161861 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.162069 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.162144 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.161998 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.167900 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg"] Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.318738 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh6fv\" (UniqueName: \"kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.318839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.318861 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.318879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.421142 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.421204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.421239 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.421427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6fv\" (UniqueName: \"kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.426768 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.429453 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.430155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.438339 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh6fv\" (UniqueName: \"kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.486786 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:43:57 crc kubenswrapper[4754]: I1011 03:43:57.975239 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg"] Oct 11 03:43:57 crc kubenswrapper[4754]: W1011 03:43:57.983208 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8f433cb_12ea_454a_8f71_61d75926dcc2.slice/crio-e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f WatchSource:0}: Error finding container e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f: Status 404 returned error can't find the container with id e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f Oct 11 03:43:58 crc kubenswrapper[4754]: I1011 03:43:58.055617 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" event={"ID":"f8f433cb-12ea-454a-8f71-61d75926dcc2","Type":"ContainerStarted","Data":"e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f"} Oct 11 03:43:59 crc kubenswrapper[4754]: I1011 03:43:59.064884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" event={"ID":"f8f433cb-12ea-454a-8f71-61d75926dcc2","Type":"ContainerStarted","Data":"ae623ab7795673d5b0e3b040d537b194aa32d813699c88e2e37e2896da582dde"} Oct 11 03:43:59 crc kubenswrapper[4754]: I1011 03:43:59.084492 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" podStartSLOduration=1.463371329 podStartE2EDuration="2.084478711s" podCreationTimestamp="2025-10-11 03:43:57 +0000 UTC" firstStartedPulling="2025-10-11 03:43:57.985615471 +0000 UTC m=+2285.544560256" lastFinishedPulling="2025-10-11 03:43:58.606722853 +0000 UTC m=+2286.165667638" observedRunningTime="2025-10-11 03:43:59.077488747 +0000 UTC m=+2286.636433572" watchObservedRunningTime="2025-10-11 03:43:59.084478711 +0000 UTC m=+2286.643423496" Oct 11 03:44:09 crc kubenswrapper[4754]: I1011 03:44:09.148370 4754 generic.go:334] "Generic (PLEG): container finished" podID="f8f433cb-12ea-454a-8f71-61d75926dcc2" containerID="ae623ab7795673d5b0e3b040d537b194aa32d813699c88e2e37e2896da582dde" exitCode=0 Oct 11 03:44:09 crc kubenswrapper[4754]: I1011 03:44:09.148467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" event={"ID":"f8f433cb-12ea-454a-8f71-61d75926dcc2","Type":"ContainerDied","Data":"ae623ab7795673d5b0e3b040d537b194aa32d813699c88e2e37e2896da582dde"} Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.538424 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.560892 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph\") pod \"f8f433cb-12ea-454a-8f71-61d75926dcc2\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.560978 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh6fv\" (UniqueName: \"kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv\") pod \"f8f433cb-12ea-454a-8f71-61d75926dcc2\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.561030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key\") pod \"f8f433cb-12ea-454a-8f71-61d75926dcc2\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.561054 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory\") pod \"f8f433cb-12ea-454a-8f71-61d75926dcc2\" (UID: \"f8f433cb-12ea-454a-8f71-61d75926dcc2\") " Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.567037 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph" (OuterVolumeSpecName: "ceph") pod "f8f433cb-12ea-454a-8f71-61d75926dcc2" (UID: "f8f433cb-12ea-454a-8f71-61d75926dcc2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.576195 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv" (OuterVolumeSpecName: "kube-api-access-rh6fv") pod "f8f433cb-12ea-454a-8f71-61d75926dcc2" (UID: "f8f433cb-12ea-454a-8f71-61d75926dcc2"). InnerVolumeSpecName "kube-api-access-rh6fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.601105 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory" (OuterVolumeSpecName: "inventory") pod "f8f433cb-12ea-454a-8f71-61d75926dcc2" (UID: "f8f433cb-12ea-454a-8f71-61d75926dcc2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.602124 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8f433cb-12ea-454a-8f71-61d75926dcc2" (UID: "f8f433cb-12ea-454a-8f71-61d75926dcc2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.663696 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.663730 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh6fv\" (UniqueName: \"kubernetes.io/projected/f8f433cb-12ea-454a-8f71-61d75926dcc2-kube-api-access-rh6fv\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.663739 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:10 crc kubenswrapper[4754]: I1011 03:44:10.663747 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f433cb-12ea-454a-8f71-61d75926dcc2-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.175705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" event={"ID":"f8f433cb-12ea-454a-8f71-61d75926dcc2","Type":"ContainerDied","Data":"e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f"} Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.175767 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68e9cd6e5e63b1410da5b9d0ab83256f25fae337870558b674914ba12e6813f" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.175876 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.247129 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8"] Oct 11 03:44:11 crc kubenswrapper[4754]: E1011 03:44:11.247555 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f433cb-12ea-454a-8f71-61d75926dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.247574 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f433cb-12ea-454a-8f71-61d75926dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.247811 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f433cb-12ea-454a-8f71-61d75926dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.248841 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.253152 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.253467 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.253807 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.253865 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.253985 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.254655 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.257134 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.257654 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8"] Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.262327 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.275914 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276091 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276145 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276192 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276225 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276266 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62wgk\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276409 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276445 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276486 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.276536 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377417 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62wgk\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377764 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377839 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377908 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.377949 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.378001 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.378043 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.378086 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.378124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.378150 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.383238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.383502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.383560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.384561 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.384759 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.384784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.384894 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.385849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.386954 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.388036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.388528 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.391556 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.396443 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62wgk\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:11 crc kubenswrapper[4754]: I1011 03:44:11.579658 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:12 crc kubenswrapper[4754]: I1011 03:44:12.091412 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8"] Oct 11 03:44:12 crc kubenswrapper[4754]: I1011 03:44:12.183458 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" event={"ID":"4533b55f-2b04-4059-9ed9-f5be54f606b3","Type":"ContainerStarted","Data":"952492ecfdbc3d652dbb1bdf02764d04d479bdf529f7321076111e5dffd26ccb"} Oct 11 03:44:13 crc kubenswrapper[4754]: I1011 03:44:13.194799 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" event={"ID":"4533b55f-2b04-4059-9ed9-f5be54f606b3","Type":"ContainerStarted","Data":"659ae825e9e733a85a3223582103c8dc620a5e02326de58ea43e74d4bd38cb95"} Oct 11 03:44:13 crc kubenswrapper[4754]: I1011 03:44:13.220385 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" podStartSLOduration=1.821929719 podStartE2EDuration="2.220361647s" podCreationTimestamp="2025-10-11 03:44:11 +0000 UTC" firstStartedPulling="2025-10-11 03:44:12.090346834 +0000 UTC m=+2299.649291619" lastFinishedPulling="2025-10-11 03:44:12.488778752 +0000 UTC m=+2300.047723547" observedRunningTime="2025-10-11 03:44:13.214644539 +0000 UTC m=+2300.773589334" watchObservedRunningTime="2025-10-11 03:44:13.220361647 +0000 UTC m=+2300.779306452" Oct 11 03:44:30 crc kubenswrapper[4754]: I1011 03:44:30.736866 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:44:30 crc kubenswrapper[4754]: I1011 03:44:30.737452 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:44:48 crc kubenswrapper[4754]: I1011 03:44:48.518176 4754 generic.go:334] "Generic (PLEG): container finished" podID="4533b55f-2b04-4059-9ed9-f5be54f606b3" containerID="659ae825e9e733a85a3223582103c8dc620a5e02326de58ea43e74d4bd38cb95" exitCode=0 Oct 11 03:44:48 crc kubenswrapper[4754]: I1011 03:44:48.518306 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" event={"ID":"4533b55f-2b04-4059-9ed9-f5be54f606b3","Type":"ContainerDied","Data":"659ae825e9e733a85a3223582103c8dc620a5e02326de58ea43e74d4bd38cb95"} Oct 11 03:44:49 crc kubenswrapper[4754]: I1011 03:44:49.965308 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.047872 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048025 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048079 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62wgk\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048259 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048323 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048388 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048468 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048504 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048575 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.048615 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key\") pod \"4533b55f-2b04-4059-9ed9-f5be54f606b3\" (UID: \"4533b55f-2b04-4059-9ed9-f5be54f606b3\") " Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.056333 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.056412 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.056517 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.056593 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.057693 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk" (OuterVolumeSpecName: "kube-api-access-62wgk") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "kube-api-access-62wgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.059247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.059368 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph" (OuterVolumeSpecName: "ceph") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.060320 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.060552 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.060879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.066101 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.078279 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.079367 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory" (OuterVolumeSpecName: "inventory") pod "4533b55f-2b04-4059-9ed9-f5be54f606b3" (UID: "4533b55f-2b04-4059-9ed9-f5be54f606b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.150763 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151008 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151146 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62wgk\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-kube-api-access-62wgk\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151292 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151456 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151571 4754 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151677 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151795 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.151911 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4533b55f-2b04-4059-9ed9-f5be54f606b3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.152061 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.152171 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.152284 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.152400 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4533b55f-2b04-4059-9ed9-f5be54f606b3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.543184 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" event={"ID":"4533b55f-2b04-4059-9ed9-f5be54f606b3","Type":"ContainerDied","Data":"952492ecfdbc3d652dbb1bdf02764d04d479bdf529f7321076111e5dffd26ccb"} Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.543443 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="952492ecfdbc3d652dbb1bdf02764d04d479bdf529f7321076111e5dffd26ccb" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.543287 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.661730 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz"] Oct 11 03:44:50 crc kubenswrapper[4754]: E1011 03:44:50.662078 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4533b55f-2b04-4059-9ed9-f5be54f606b3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.662095 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4533b55f-2b04-4059-9ed9-f5be54f606b3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.662311 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4533b55f-2b04-4059-9ed9-f5be54f606b3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.662989 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.672606 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.672939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfcgr\" (UniqueName: \"kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.673063 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.673098 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.675285 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz"] Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.716251 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.716343 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.716478 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.716589 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.716670 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.774879 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.775000 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfcgr\" (UniqueName: \"kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.775132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.775177 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.779783 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.782450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.782475 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:50 crc kubenswrapper[4754]: I1011 03:44:50.792984 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfcgr\" (UniqueName: \"kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:51 crc kubenswrapper[4754]: I1011 03:44:51.039350 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:44:51 crc kubenswrapper[4754]: I1011 03:44:51.562910 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz"] Oct 11 03:44:51 crc kubenswrapper[4754]: I1011 03:44:51.569003 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:44:52 crc kubenswrapper[4754]: I1011 03:44:52.568538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" event={"ID":"986f6f39-c174-4f1e-a798-6ee6c389245e","Type":"ContainerStarted","Data":"314975ae543d60778dd14414d7813322ef299831d155ce4a3b6cccd786ae518b"} Oct 11 03:44:52 crc kubenswrapper[4754]: I1011 03:44:52.568878 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" event={"ID":"986f6f39-c174-4f1e-a798-6ee6c389245e","Type":"ContainerStarted","Data":"7d1d646eb7ec541177438f5a94fabfcc51f192aa5fecffd318f742a2a670af52"} Oct 11 03:44:52 crc kubenswrapper[4754]: I1011 03:44:52.597279 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" podStartSLOduration=2.097883672 podStartE2EDuration="2.597253799s" podCreationTimestamp="2025-10-11 03:44:50 +0000 UTC" firstStartedPulling="2025-10-11 03:44:51.568711229 +0000 UTC m=+2339.127656014" lastFinishedPulling="2025-10-11 03:44:52.068081356 +0000 UTC m=+2339.627026141" observedRunningTime="2025-10-11 03:44:52.584600488 +0000 UTC m=+2340.143545273" watchObservedRunningTime="2025-10-11 03:44:52.597253799 +0000 UTC m=+2340.156198614" Oct 11 03:44:58 crc kubenswrapper[4754]: I1011 03:44:58.615763 4754 generic.go:334] "Generic (PLEG): container finished" podID="986f6f39-c174-4f1e-a798-6ee6c389245e" containerID="314975ae543d60778dd14414d7813322ef299831d155ce4a3b6cccd786ae518b" exitCode=0 Oct 11 03:44:58 crc kubenswrapper[4754]: I1011 03:44:58.615845 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" event={"ID":"986f6f39-c174-4f1e-a798-6ee6c389245e","Type":"ContainerDied","Data":"314975ae543d60778dd14414d7813322ef299831d155ce4a3b6cccd786ae518b"} Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.011817 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.137889 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm"] Oct 11 03:45:00 crc kubenswrapper[4754]: E1011 03:45:00.138314 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f6f39-c174-4f1e-a798-6ee6c389245e" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.138330 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f6f39-c174-4f1e-a798-6ee6c389245e" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.138551 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f6f39-c174-4f1e-a798-6ee6c389245e" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.139181 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.141470 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.141829 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.146872 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm"] Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.180391 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") pod \"986f6f39-c174-4f1e-a798-6ee6c389245e\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.180621 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph\") pod \"986f6f39-c174-4f1e-a798-6ee6c389245e\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.180676 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfcgr\" (UniqueName: \"kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr\") pod \"986f6f39-c174-4f1e-a798-6ee6c389245e\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.180731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key\") pod \"986f6f39-c174-4f1e-a798-6ee6c389245e\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.191789 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph" (OuterVolumeSpecName: "ceph") pod "986f6f39-c174-4f1e-a798-6ee6c389245e" (UID: "986f6f39-c174-4f1e-a798-6ee6c389245e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.191932 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr" (OuterVolumeSpecName: "kube-api-access-cfcgr") pod "986f6f39-c174-4f1e-a798-6ee6c389245e" (UID: "986f6f39-c174-4f1e-a798-6ee6c389245e"). InnerVolumeSpecName "kube-api-access-cfcgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:45:00 crc kubenswrapper[4754]: E1011 03:45:00.206227 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory podName:986f6f39-c174-4f1e-a798-6ee6c389245e nodeName:}" failed. No retries permitted until 2025-10-11 03:45:00.706198949 +0000 UTC m=+2348.265143734 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory") pod "986f6f39-c174-4f1e-a798-6ee6c389245e" (UID: "986f6f39-c174-4f1e-a798-6ee6c389245e") : error deleting /var/lib/kubelet/pods/986f6f39-c174-4f1e-a798-6ee6c389245e/volume-subpaths: remove /var/lib/kubelet/pods/986f6f39-c174-4f1e-a798-6ee6c389245e/volume-subpaths: no such file or directory Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.208731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "986f6f39-c174-4f1e-a798-6ee6c389245e" (UID: "986f6f39-c174-4f1e-a798-6ee6c389245e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.283027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.283286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfknn\" (UniqueName: \"kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.283420 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.284245 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.284269 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfcgr\" (UniqueName: \"kubernetes.io/projected/986f6f39-c174-4f1e-a798-6ee6c389245e-kube-api-access-cfcgr\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.284278 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.386043 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.386085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfknn\" (UniqueName: \"kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.386147 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.386986 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.389984 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.403270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfknn\" (UniqueName: \"kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn\") pod \"collect-profiles-29335905-ttmkm\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.460769 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.636192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" event={"ID":"986f6f39-c174-4f1e-a798-6ee6c389245e","Type":"ContainerDied","Data":"7d1d646eb7ec541177438f5a94fabfcc51f192aa5fecffd318f742a2a670af52"} Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.636235 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d1d646eb7ec541177438f5a94fabfcc51f192aa5fecffd318f742a2a670af52" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.636249 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.704233 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc"] Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.705694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.710510 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.718886 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc"] Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.735948 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.736065 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.796952 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") pod \"986f6f39-c174-4f1e-a798-6ee6c389245e\" (UID: \"986f6f39-c174-4f1e-a798-6ee6c389245e\") " Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.801638 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory" (OuterVolumeSpecName: "inventory") pod "986f6f39-c174-4f1e-a798-6ee6c389245e" (UID: "986f6f39-c174-4f1e-a798-6ee6c389245e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.860736 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm"] Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.899390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlm8p\" (UniqueName: \"kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.899462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.899684 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.899779 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.899856 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.900064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:00 crc kubenswrapper[4754]: I1011 03:45:00.900304 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986f6f39-c174-4f1e-a798-6ee6c389245e-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001593 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001626 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001674 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlm8p\" (UniqueName: \"kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.001777 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.002570 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.008725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.009158 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.009940 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.010383 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.018648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlm8p\" (UniqueName: \"kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hxntc\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.027531 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.507126 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc"] Oct 11 03:45:01 crc kubenswrapper[4754]: W1011 03:45:01.511104 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51e593be_a82a_499d_91fd_f2e05d771519.slice/crio-b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832 WatchSource:0}: Error finding container b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832: Status 404 returned error can't find the container with id b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832 Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.643409 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" event={"ID":"51e593be-a82a-499d-91fd-f2e05d771519","Type":"ContainerStarted","Data":"b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832"} Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.645336 4754 generic.go:334] "Generic (PLEG): container finished" podID="a09745f4-c2b2-45b0-98f8-23779ee4144e" containerID="1307327f11d36ec0cd018b1643c7192cbbe028e37a1a1cfa6e0928d63de5c4c9" exitCode=0 Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.645361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" event={"ID":"a09745f4-c2b2-45b0-98f8-23779ee4144e","Type":"ContainerDied","Data":"1307327f11d36ec0cd018b1643c7192cbbe028e37a1a1cfa6e0928d63de5c4c9"} Oct 11 03:45:01 crc kubenswrapper[4754]: I1011 03:45:01.645376 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" event={"ID":"a09745f4-c2b2-45b0-98f8-23779ee4144e","Type":"ContainerStarted","Data":"aab2461cdfebe5f1e089bd92e1162904c301a7f062833324894fad0c0310805d"} Oct 11 03:45:02 crc kubenswrapper[4754]: I1011 03:45:02.657038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" event={"ID":"51e593be-a82a-499d-91fd-f2e05d771519","Type":"ContainerStarted","Data":"fc965b149340f547352a64dedf0d3423e5d01c1a6483ccd968f0d8c5095a0a50"} Oct 11 03:45:02 crc kubenswrapper[4754]: I1011 03:45:02.682712 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" podStartSLOduration=2.254392121 podStartE2EDuration="2.682691976s" podCreationTimestamp="2025-10-11 03:45:00 +0000 UTC" firstStartedPulling="2025-10-11 03:45:01.51322918 +0000 UTC m=+2349.072173965" lastFinishedPulling="2025-10-11 03:45:01.941529035 +0000 UTC m=+2349.500473820" observedRunningTime="2025-10-11 03:45:02.675591429 +0000 UTC m=+2350.234536254" watchObservedRunningTime="2025-10-11 03:45:02.682691976 +0000 UTC m=+2350.241636761" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.066574 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.240297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfknn\" (UniqueName: \"kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn\") pod \"a09745f4-c2b2-45b0-98f8-23779ee4144e\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.240370 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume\") pod \"a09745f4-c2b2-45b0-98f8-23779ee4144e\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.240511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume\") pod \"a09745f4-c2b2-45b0-98f8-23779ee4144e\" (UID: \"a09745f4-c2b2-45b0-98f8-23779ee4144e\") " Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.241404 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume" (OuterVolumeSpecName: "config-volume") pod "a09745f4-c2b2-45b0-98f8-23779ee4144e" (UID: "a09745f4-c2b2-45b0-98f8-23779ee4144e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.245799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a09745f4-c2b2-45b0-98f8-23779ee4144e" (UID: "a09745f4-c2b2-45b0-98f8-23779ee4144e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.246166 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn" (OuterVolumeSpecName: "kube-api-access-zfknn") pod "a09745f4-c2b2-45b0-98f8-23779ee4144e" (UID: "a09745f4-c2b2-45b0-98f8-23779ee4144e"). InnerVolumeSpecName "kube-api-access-zfknn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.342636 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfknn\" (UniqueName: \"kubernetes.io/projected/a09745f4-c2b2-45b0-98f8-23779ee4144e-kube-api-access-zfknn\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.342666 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a09745f4-c2b2-45b0-98f8-23779ee4144e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.342674 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a09745f4-c2b2-45b0-98f8-23779ee4144e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.667230 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.672145 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm" event={"ID":"a09745f4-c2b2-45b0-98f8-23779ee4144e","Type":"ContainerDied","Data":"aab2461cdfebe5f1e089bd92e1162904c301a7f062833324894fad0c0310805d"} Oct 11 03:45:03 crc kubenswrapper[4754]: I1011 03:45:03.672203 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aab2461cdfebe5f1e089bd92e1162904c301a7f062833324894fad0c0310805d" Oct 11 03:45:04 crc kubenswrapper[4754]: I1011 03:45:04.133193 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw"] Oct 11 03:45:04 crc kubenswrapper[4754]: I1011 03:45:04.139270 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335860-pv7qw"] Oct 11 03:45:05 crc kubenswrapper[4754]: I1011 03:45:05.096159 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afd23974-0ea6-466b-a570-2d1d32c12a84" path="/var/lib/kubelet/pods/afd23974-0ea6-466b-a570-2d1d32c12a84/volumes" Oct 11 03:45:06 crc kubenswrapper[4754]: I1011 03:45:06.058228 4754 scope.go:117] "RemoveContainer" containerID="703a85e9411fe67d8e4afb6675677e8ea68737ffdb850166aaf66a1b37bb99d1" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.736174 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.736670 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.736718 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.737465 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.737519 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" gracePeriod=600 Oct 11 03:45:30 crc kubenswrapper[4754]: E1011 03:45:30.863667 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.883896 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" exitCode=0 Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.883938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6"} Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.883979 4754 scope.go:117] "RemoveContainer" containerID="06ad59408c120a3b38af0a46ded5787ac4355a10743af621b51f2b8388c0ca9a" Oct 11 03:45:30 crc kubenswrapper[4754]: I1011 03:45:30.884621 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:45:30 crc kubenswrapper[4754]: E1011 03:45:30.884913 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:45:44 crc kubenswrapper[4754]: I1011 03:45:44.083667 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:45:44 crc kubenswrapper[4754]: E1011 03:45:44.084383 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:45:56 crc kubenswrapper[4754]: I1011 03:45:56.083661 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:45:56 crc kubenswrapper[4754]: E1011 03:45:56.084457 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:46:11 crc kubenswrapper[4754]: I1011 03:46:11.083711 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:46:11 crc kubenswrapper[4754]: E1011 03:46:11.084455 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:46:18 crc kubenswrapper[4754]: I1011 03:46:18.285089 4754 generic.go:334] "Generic (PLEG): container finished" podID="51e593be-a82a-499d-91fd-f2e05d771519" containerID="fc965b149340f547352a64dedf0d3423e5d01c1a6483ccd968f0d8c5095a0a50" exitCode=0 Oct 11 03:46:18 crc kubenswrapper[4754]: I1011 03:46:18.285337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" event={"ID":"51e593be-a82a-499d-91fd-f2e05d771519","Type":"ContainerDied","Data":"fc965b149340f547352a64dedf0d3423e5d01c1a6483ccd968f0d8c5095a0a50"} Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.659309 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.842936 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlm8p\" (UniqueName: \"kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.843269 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.843316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.843371 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.843412 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.843462 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory\") pod \"51e593be-a82a-499d-91fd-f2e05d771519\" (UID: \"51e593be-a82a-499d-91fd-f2e05d771519\") " Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.849125 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.849372 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph" (OuterVolumeSpecName: "ceph") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.849509 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p" (OuterVolumeSpecName: "kube-api-access-xlm8p") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "kube-api-access-xlm8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.867913 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.872799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.894582 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory" (OuterVolumeSpecName: "inventory") pod "51e593be-a82a-499d-91fd-f2e05d771519" (UID: "51e593be-a82a-499d-91fd-f2e05d771519"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946220 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946274 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946286 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946300 4754 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51e593be-a82a-499d-91fd-f2e05d771519-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946327 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e593be-a82a-499d-91fd-f2e05d771519-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:19 crc kubenswrapper[4754]: I1011 03:46:19.946339 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlm8p\" (UniqueName: \"kubernetes.io/projected/51e593be-a82a-499d-91fd-f2e05d771519-kube-api-access-xlm8p\") on node \"crc\" DevicePath \"\"" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.307307 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" event={"ID":"51e593be-a82a-499d-91fd-f2e05d771519","Type":"ContainerDied","Data":"b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832"} Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.307348 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b682faba561050d2416dcd5cf1e0ca88c617f07643a9cfe31d17e47f5e9f6832" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.307412 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hxntc" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.422367 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg"] Oct 11 03:46:20 crc kubenswrapper[4754]: E1011 03:46:20.422729 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e593be-a82a-499d-91fd-f2e05d771519" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.422746 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e593be-a82a-499d-91fd-f2e05d771519" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 11 03:46:20 crc kubenswrapper[4754]: E1011 03:46:20.422763 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a09745f4-c2b2-45b0-98f8-23779ee4144e" containerName="collect-profiles" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.422770 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a09745f4-c2b2-45b0-98f8-23779ee4144e" containerName="collect-profiles" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.422944 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a09745f4-c2b2-45b0-98f8-23779ee4144e" containerName="collect-profiles" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.423062 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e593be-a82a-499d-91fd-f2e05d771519" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.423705 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.428677 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.429288 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.429445 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.429492 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.429930 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.430083 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.430553 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.434529 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg"] Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456607 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wllvq\" (UniqueName: \"kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456656 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456724 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.456845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558284 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558339 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558373 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wllvq\" (UniqueName: \"kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558412 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558443 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.558459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.562401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.562424 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.562798 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.563074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.563104 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.563918 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.575689 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wllvq\" (UniqueName: \"kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:20 crc kubenswrapper[4754]: I1011 03:46:20.739979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:46:21 crc kubenswrapper[4754]: I1011 03:46:21.221765 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg"] Oct 11 03:46:21 crc kubenswrapper[4754]: I1011 03:46:21.318467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" event={"ID":"e12f53c3-0230-41c0-a33a-d59605640584","Type":"ContainerStarted","Data":"d2ed27ea7b7e71e324c93e2dd2884f916679eb681b14d771ea72d18202057bbf"} Oct 11 03:46:23 crc kubenswrapper[4754]: I1011 03:46:23.093175 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:46:23 crc kubenswrapper[4754]: E1011 03:46:23.093917 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:46:23 crc kubenswrapper[4754]: I1011 03:46:23.333858 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" event={"ID":"e12f53c3-0230-41c0-a33a-d59605640584","Type":"ContainerStarted","Data":"c9c8e4efa9b6efdb8bbd6f71a0687e4779769aa50c353de3f2f76a4cc81e9a74"} Oct 11 03:46:23 crc kubenswrapper[4754]: I1011 03:46:23.356717 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" podStartSLOduration=2.219156121 podStartE2EDuration="3.356695815s" podCreationTimestamp="2025-10-11 03:46:20 +0000 UTC" firstStartedPulling="2025-10-11 03:46:21.24226479 +0000 UTC m=+2428.801209585" lastFinishedPulling="2025-10-11 03:46:22.379804494 +0000 UTC m=+2429.938749279" observedRunningTime="2025-10-11 03:46:23.35471429 +0000 UTC m=+2430.913659095" watchObservedRunningTime="2025-10-11 03:46:23.356695815 +0000 UTC m=+2430.915640600" Oct 11 03:46:35 crc kubenswrapper[4754]: I1011 03:46:35.084107 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:46:35 crc kubenswrapper[4754]: E1011 03:46:35.084955 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:46:48 crc kubenswrapper[4754]: I1011 03:46:48.083546 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:46:48 crc kubenswrapper[4754]: E1011 03:46:48.084466 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:46:59 crc kubenswrapper[4754]: I1011 03:46:59.086320 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:46:59 crc kubenswrapper[4754]: E1011 03:46:59.087007 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:47:14 crc kubenswrapper[4754]: I1011 03:47:14.083899 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:47:14 crc kubenswrapper[4754]: E1011 03:47:14.085172 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:47:22 crc kubenswrapper[4754]: I1011 03:47:22.856449 4754 generic.go:334] "Generic (PLEG): container finished" podID="e12f53c3-0230-41c0-a33a-d59605640584" containerID="c9c8e4efa9b6efdb8bbd6f71a0687e4779769aa50c353de3f2f76a4cc81e9a74" exitCode=0 Oct 11 03:47:22 crc kubenswrapper[4754]: I1011 03:47:22.856545 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" event={"ID":"e12f53c3-0230-41c0-a33a-d59605640584","Type":"ContainerDied","Data":"c9c8e4efa9b6efdb8bbd6f71a0687e4779769aa50c353de3f2f76a4cc81e9a74"} Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.283591 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428236 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428299 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428379 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428430 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428478 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wllvq\" (UniqueName: \"kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428503 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.428546 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0\") pod \"e12f53c3-0230-41c0-a33a-d59605640584\" (UID: \"e12f53c3-0230-41c0-a33a-d59605640584\") " Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.435188 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq" (OuterVolumeSpecName: "kube-api-access-wllvq") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "kube-api-access-wllvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.435600 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph" (OuterVolumeSpecName: "ceph") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.447307 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.459713 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory" (OuterVolumeSpecName: "inventory") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.461683 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.463287 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.472609 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e12f53c3-0230-41c0-a33a-d59605640584" (UID: "e12f53c3-0230-41c0-a33a-d59605640584"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534048 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wllvq\" (UniqueName: \"kubernetes.io/projected/e12f53c3-0230-41c0-a33a-d59605640584-kube-api-access-wllvq\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534089 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534102 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534140 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534152 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534162 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.534170 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e12f53c3-0230-41c0-a33a-d59605640584-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.874890 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" event={"ID":"e12f53c3-0230-41c0-a33a-d59605640584","Type":"ContainerDied","Data":"d2ed27ea7b7e71e324c93e2dd2884f916679eb681b14d771ea72d18202057bbf"} Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.874938 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2ed27ea7b7e71e324c93e2dd2884f916679eb681b14d771ea72d18202057bbf" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.874949 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.969894 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc"] Oct 11 03:47:24 crc kubenswrapper[4754]: E1011 03:47:24.970393 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f53c3-0230-41c0-a33a-d59605640584" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.970418 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f53c3-0230-41c0-a33a-d59605640584" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.970649 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12f53c3-0230-41c0-a33a-d59605640584" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.971374 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974079 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974261 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974448 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974570 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974573 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.974706 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:47:24 crc kubenswrapper[4754]: I1011 03:47:24.979359 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc"] Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.084525 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:47:25 crc kubenswrapper[4754]: E1011 03:47:25.084830 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143715 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143742 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143781 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4pk5\" (UniqueName: \"kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143802 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.143864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245481 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4pk5\" (UniqueName: \"kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245513 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245571 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.245779 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.249637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.263043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.268023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.268986 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.272654 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4pk5\" (UniqueName: \"kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.276183 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b59cc\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.287488 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.780704 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc"] Oct 11 03:47:25 crc kubenswrapper[4754]: I1011 03:47:25.884097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" event={"ID":"347dff33-1661-4ed0-97b0-4248e8648f6a","Type":"ContainerStarted","Data":"bb54b6761ec166e0966ecc22aca3620b289772ef5c967d41f3356879ab62ecf6"} Oct 11 03:47:26 crc kubenswrapper[4754]: I1011 03:47:26.893044 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" event={"ID":"347dff33-1661-4ed0-97b0-4248e8648f6a","Type":"ContainerStarted","Data":"4c2f54d27095a3e964c95c324e2255ee004bef1dea5493d5e71af2ba7be3410f"} Oct 11 03:47:26 crc kubenswrapper[4754]: I1011 03:47:26.919202 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" podStartSLOduration=2.17915135 podStartE2EDuration="2.919184607s" podCreationTimestamp="2025-10-11 03:47:24 +0000 UTC" firstStartedPulling="2025-10-11 03:47:25.789913879 +0000 UTC m=+2493.348858664" lastFinishedPulling="2025-10-11 03:47:26.529947146 +0000 UTC m=+2494.088891921" observedRunningTime="2025-10-11 03:47:26.911755093 +0000 UTC m=+2494.470699878" watchObservedRunningTime="2025-10-11 03:47:26.919184607 +0000 UTC m=+2494.478129392" Oct 11 03:47:37 crc kubenswrapper[4754]: I1011 03:47:37.083768 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:47:37 crc kubenswrapper[4754]: E1011 03:47:37.084771 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:47:50 crc kubenswrapper[4754]: I1011 03:47:50.085467 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:47:50 crc kubenswrapper[4754]: E1011 03:47:50.087034 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:48:04 crc kubenswrapper[4754]: I1011 03:48:04.084246 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:48:04 crc kubenswrapper[4754]: E1011 03:48:04.085421 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:48:16 crc kubenswrapper[4754]: I1011 03:48:16.084354 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:48:16 crc kubenswrapper[4754]: E1011 03:48:16.085398 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:48:27 crc kubenswrapper[4754]: I1011 03:48:27.084301 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:48:27 crc kubenswrapper[4754]: E1011 03:48:27.085181 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:48:42 crc kubenswrapper[4754]: I1011 03:48:42.084120 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:48:42 crc kubenswrapper[4754]: E1011 03:48:42.085129 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:48:53 crc kubenswrapper[4754]: I1011 03:48:53.092787 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:48:53 crc kubenswrapper[4754]: E1011 03:48:53.094017 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:49:04 crc kubenswrapper[4754]: I1011 03:49:04.083345 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:49:04 crc kubenswrapper[4754]: E1011 03:49:04.084407 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:49:15 crc kubenswrapper[4754]: I1011 03:49:15.084107 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:49:15 crc kubenswrapper[4754]: E1011 03:49:15.084929 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:49:29 crc kubenswrapper[4754]: I1011 03:49:29.084316 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:49:29 crc kubenswrapper[4754]: E1011 03:49:29.085111 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:49:43 crc kubenswrapper[4754]: I1011 03:49:43.096299 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:49:43 crc kubenswrapper[4754]: E1011 03:49:43.097309 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:49:54 crc kubenswrapper[4754]: I1011 03:49:54.085695 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:49:54 crc kubenswrapper[4754]: E1011 03:49:54.087238 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:50:06 crc kubenswrapper[4754]: I1011 03:50:06.083594 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:50:06 crc kubenswrapper[4754]: E1011 03:50:06.084587 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:50:20 crc kubenswrapper[4754]: I1011 03:50:20.083747 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:50:20 crc kubenswrapper[4754]: E1011 03:50:20.084608 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:50:35 crc kubenswrapper[4754]: I1011 03:50:35.084910 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:50:35 crc kubenswrapper[4754]: I1011 03:50:35.663782 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4"} Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.129466 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.136122 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.150240 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.269553 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw2dd\" (UniqueName: \"kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.269682 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.269706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.371279 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.371335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.371429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw2dd\" (UniqueName: \"kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.372399 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.372399 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.393255 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw2dd\" (UniqueName: \"kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd\") pod \"redhat-operators-rtr2h\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.461394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:19 crc kubenswrapper[4754]: I1011 03:51:19.929716 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:19 crc kubenswrapper[4754]: W1011 03:51:19.932746 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b55f39_8ab8_4f6d_854e_e0f93c204251.slice/crio-af5563f18fe763c8be38c053afeb7a385f9f65ef0d650db96e829edaac9a2fb8 WatchSource:0}: Error finding container af5563f18fe763c8be38c053afeb7a385f9f65ef0d650db96e829edaac9a2fb8: Status 404 returned error can't find the container with id af5563f18fe763c8be38c053afeb7a385f9f65ef0d650db96e829edaac9a2fb8 Oct 11 03:51:20 crc kubenswrapper[4754]: I1011 03:51:20.155504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerStarted","Data":"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20"} Oct 11 03:51:20 crc kubenswrapper[4754]: I1011 03:51:20.155918 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerStarted","Data":"af5563f18fe763c8be38c053afeb7a385f9f65ef0d650db96e829edaac9a2fb8"} Oct 11 03:51:21 crc kubenswrapper[4754]: I1011 03:51:21.183046 4754 generic.go:334] "Generic (PLEG): container finished" podID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerID="a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20" exitCode=0 Oct 11 03:51:21 crc kubenswrapper[4754]: I1011 03:51:21.183199 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerDied","Data":"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20"} Oct 11 03:51:21 crc kubenswrapper[4754]: I1011 03:51:21.186999 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:51:22 crc kubenswrapper[4754]: I1011 03:51:22.193865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerStarted","Data":"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87"} Oct 11 03:51:23 crc kubenswrapper[4754]: I1011 03:51:23.206588 4754 generic.go:334] "Generic (PLEG): container finished" podID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerID="cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87" exitCode=0 Oct 11 03:51:23 crc kubenswrapper[4754]: I1011 03:51:23.207264 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerDied","Data":"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87"} Oct 11 03:51:24 crc kubenswrapper[4754]: I1011 03:51:24.222740 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerStarted","Data":"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50"} Oct 11 03:51:24 crc kubenswrapper[4754]: I1011 03:51:24.257517 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rtr2h" podStartSLOduration=2.8385663450000003 podStartE2EDuration="5.2574713s" podCreationTimestamp="2025-10-11 03:51:19 +0000 UTC" firstStartedPulling="2025-10-11 03:51:21.186700742 +0000 UTC m=+2728.745645527" lastFinishedPulling="2025-10-11 03:51:23.605605697 +0000 UTC m=+2731.164550482" observedRunningTime="2025-10-11 03:51:24.240932004 +0000 UTC m=+2731.799876819" watchObservedRunningTime="2025-10-11 03:51:24.2574713 +0000 UTC m=+2731.816416135" Oct 11 03:51:29 crc kubenswrapper[4754]: I1011 03:51:29.462069 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:29 crc kubenswrapper[4754]: I1011 03:51:29.462567 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:29 crc kubenswrapper[4754]: I1011 03:51:29.537221 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:30 crc kubenswrapper[4754]: I1011 03:51:30.358172 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:30 crc kubenswrapper[4754]: I1011 03:51:30.427207 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.311674 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rtr2h" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="registry-server" containerID="cri-o://c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50" gracePeriod=2 Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.749636 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.849870 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities\") pod \"66b55f39-8ab8-4f6d-854e-e0f93c204251\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.850368 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content\") pod \"66b55f39-8ab8-4f6d-854e-e0f93c204251\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.850446 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw2dd\" (UniqueName: \"kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd\") pod \"66b55f39-8ab8-4f6d-854e-e0f93c204251\" (UID: \"66b55f39-8ab8-4f6d-854e-e0f93c204251\") " Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.851022 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities" (OuterVolumeSpecName: "utilities") pod "66b55f39-8ab8-4f6d-854e-e0f93c204251" (UID: "66b55f39-8ab8-4f6d-854e-e0f93c204251"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.857096 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd" (OuterVolumeSpecName: "kube-api-access-jw2dd") pod "66b55f39-8ab8-4f6d-854e-e0f93c204251" (UID: "66b55f39-8ab8-4f6d-854e-e0f93c204251"). InnerVolumeSpecName "kube-api-access-jw2dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.952678 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4754]: I1011 03:51:32.952710 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw2dd\" (UniqueName: \"kubernetes.io/projected/66b55f39-8ab8-4f6d-854e-e0f93c204251-kube-api-access-jw2dd\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.329750 4754 generic.go:334] "Generic (PLEG): container finished" podID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerID="c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50" exitCode=0 Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.329802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerDied","Data":"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50"} Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.329844 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtr2h" event={"ID":"66b55f39-8ab8-4f6d-854e-e0f93c204251","Type":"ContainerDied","Data":"af5563f18fe763c8be38c053afeb7a385f9f65ef0d650db96e829edaac9a2fb8"} Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.329873 4754 scope.go:117] "RemoveContainer" containerID="c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.330088 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtr2h" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.367141 4754 scope.go:117] "RemoveContainer" containerID="cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.400009 4754 scope.go:117] "RemoveContainer" containerID="a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.456889 4754 scope.go:117] "RemoveContainer" containerID="c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50" Oct 11 03:51:33 crc kubenswrapper[4754]: E1011 03:51:33.457452 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50\": container with ID starting with c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50 not found: ID does not exist" containerID="c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.457517 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50"} err="failed to get container status \"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50\": rpc error: code = NotFound desc = could not find container \"c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50\": container with ID starting with c61d97380e61f1a995d35e9ec985f88da94fcc8c648083991a73f2ea32302f50 not found: ID does not exist" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.457559 4754 scope.go:117] "RemoveContainer" containerID="cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87" Oct 11 03:51:33 crc kubenswrapper[4754]: E1011 03:51:33.457991 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87\": container with ID starting with cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87 not found: ID does not exist" containerID="cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.458042 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87"} err="failed to get container status \"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87\": rpc error: code = NotFound desc = could not find container \"cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87\": container with ID starting with cd7da77df04a1640540e778821cf38203ff69694393c7588ec6e4e36cbd65f87 not found: ID does not exist" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.458080 4754 scope.go:117] "RemoveContainer" containerID="a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20" Oct 11 03:51:33 crc kubenswrapper[4754]: E1011 03:51:33.458650 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20\": container with ID starting with a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20 not found: ID does not exist" containerID="a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20" Oct 11 03:51:33 crc kubenswrapper[4754]: I1011 03:51:33.458698 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20"} err="failed to get container status \"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20\": rpc error: code = NotFound desc = could not find container \"a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20\": container with ID starting with a06540c93a2196bad62327563bb05f3d24d16c9451fdccc15d005284597b0b20 not found: ID does not exist" Oct 11 03:51:34 crc kubenswrapper[4754]: I1011 03:51:34.404531 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66b55f39-8ab8-4f6d-854e-e0f93c204251" (UID: "66b55f39-8ab8-4f6d-854e-e0f93c204251"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:34 crc kubenswrapper[4754]: I1011 03:51:34.484042 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b55f39-8ab8-4f6d-854e-e0f93c204251-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:34 crc kubenswrapper[4754]: I1011 03:51:34.606839 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:34 crc kubenswrapper[4754]: I1011 03:51:34.617806 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rtr2h"] Oct 11 03:51:35 crc kubenswrapper[4754]: I1011 03:51:35.101092 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" path="/var/lib/kubelet/pods/66b55f39-8ab8-4f6d-854e-e0f93c204251/volumes" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.192133 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:46 crc kubenswrapper[4754]: E1011 03:51:46.193659 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="extract-utilities" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.193678 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="extract-utilities" Oct 11 03:51:46 crc kubenswrapper[4754]: E1011 03:51:46.193698 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="extract-content" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.193706 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="extract-content" Oct 11 03:51:46 crc kubenswrapper[4754]: E1011 03:51:46.193722 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="registry-server" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.193731 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="registry-server" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.194026 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b55f39-8ab8-4f6d-854e-e0f93c204251" containerName="registry-server" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.198013 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.205762 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.296611 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.296694 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.296813 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tb4x\" (UniqueName: \"kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.398580 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.398641 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tb4x\" (UniqueName: \"kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.398772 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.399331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.399571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.434083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tb4x\" (UniqueName: \"kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x\") pod \"community-operators-lh4tv\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:46 crc kubenswrapper[4754]: I1011 03:51:46.533790 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:47 crc kubenswrapper[4754]: I1011 03:51:47.157255 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:47 crc kubenswrapper[4754]: I1011 03:51:47.484239 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerID="f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6" exitCode=0 Oct 11 03:51:47 crc kubenswrapper[4754]: I1011 03:51:47.484313 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerDied","Data":"f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6"} Oct 11 03:51:47 crc kubenswrapper[4754]: I1011 03:51:47.484358 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerStarted","Data":"84d148038227faf68b3586da1d141d98f767bb59566ba5db02637544bbddd9d1"} Oct 11 03:51:48 crc kubenswrapper[4754]: I1011 03:51:48.496625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerStarted","Data":"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b"} Oct 11 03:51:49 crc kubenswrapper[4754]: I1011 03:51:49.520345 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerID="b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b" exitCode=0 Oct 11 03:51:49 crc kubenswrapper[4754]: I1011 03:51:49.520465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerDied","Data":"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b"} Oct 11 03:51:50 crc kubenswrapper[4754]: I1011 03:51:50.532940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerStarted","Data":"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8"} Oct 11 03:51:50 crc kubenswrapper[4754]: I1011 03:51:50.558806 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lh4tv" podStartSLOduration=1.95020441 podStartE2EDuration="4.558777383s" podCreationTimestamp="2025-10-11 03:51:46 +0000 UTC" firstStartedPulling="2025-10-11 03:51:47.486954976 +0000 UTC m=+2755.045899781" lastFinishedPulling="2025-10-11 03:51:50.095527969 +0000 UTC m=+2757.654472754" observedRunningTime="2025-10-11 03:51:50.555613614 +0000 UTC m=+2758.114558409" watchObservedRunningTime="2025-10-11 03:51:50.558777383 +0000 UTC m=+2758.117722178" Oct 11 03:51:56 crc kubenswrapper[4754]: I1011 03:51:56.534129 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:56 crc kubenswrapper[4754]: I1011 03:51:56.535207 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:56 crc kubenswrapper[4754]: I1011 03:51:56.578221 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:56 crc kubenswrapper[4754]: I1011 03:51:56.649724 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:56 crc kubenswrapper[4754]: I1011 03:51:56.817118 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:58 crc kubenswrapper[4754]: I1011 03:51:58.608115 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lh4tv" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="registry-server" containerID="cri-o://d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8" gracePeriod=2 Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.200589 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.228159 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content\") pod \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.228577 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tb4x\" (UniqueName: \"kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x\") pod \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.228632 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities\") pod \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\" (UID: \"f6151612-e467-4803-ac1c-f10a3b7dc9b9\") " Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.230239 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities" (OuterVolumeSpecName: "utilities") pod "f6151612-e467-4803-ac1c-f10a3b7dc9b9" (UID: "f6151612-e467-4803-ac1c-f10a3b7dc9b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.236739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x" (OuterVolumeSpecName: "kube-api-access-4tb4x") pod "f6151612-e467-4803-ac1c-f10a3b7dc9b9" (UID: "f6151612-e467-4803-ac1c-f10a3b7dc9b9"). InnerVolumeSpecName "kube-api-access-4tb4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.297539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6151612-e467-4803-ac1c-f10a3b7dc9b9" (UID: "f6151612-e467-4803-ac1c-f10a3b7dc9b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.330566 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tb4x\" (UniqueName: \"kubernetes.io/projected/f6151612-e467-4803-ac1c-f10a3b7dc9b9-kube-api-access-4tb4x\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.330619 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.330629 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6151612-e467-4803-ac1c-f10a3b7dc9b9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.622210 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerID="d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8" exitCode=0 Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.622275 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerDied","Data":"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8"} Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.622325 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh4tv" event={"ID":"f6151612-e467-4803-ac1c-f10a3b7dc9b9","Type":"ContainerDied","Data":"84d148038227faf68b3586da1d141d98f767bb59566ba5db02637544bbddd9d1"} Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.622361 4754 scope.go:117] "RemoveContainer" containerID="d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.624115 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh4tv" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.677325 4754 scope.go:117] "RemoveContainer" containerID="b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.677482 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.687915 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lh4tv"] Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.711228 4754 scope.go:117] "RemoveContainer" containerID="f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.741410 4754 scope.go:117] "RemoveContainer" containerID="d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8" Oct 11 03:51:59 crc kubenswrapper[4754]: E1011 03:51:59.741998 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8\": container with ID starting with d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8 not found: ID does not exist" containerID="d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.742041 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8"} err="failed to get container status \"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8\": rpc error: code = NotFound desc = could not find container \"d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8\": container with ID starting with d9607ddd12fee6080c241cb27c23846547b2ca5831083ad55feae39e448011a8 not found: ID does not exist" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.742062 4754 scope.go:117] "RemoveContainer" containerID="b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b" Oct 11 03:51:59 crc kubenswrapper[4754]: E1011 03:51:59.742678 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b\": container with ID starting with b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b not found: ID does not exist" containerID="b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.742707 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b"} err="failed to get container status \"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b\": rpc error: code = NotFound desc = could not find container \"b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b\": container with ID starting with b3bf6245217e956172d03c38ff24488788b2b71c8871208729f513c846909d6b not found: ID does not exist" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.742726 4754 scope.go:117] "RemoveContainer" containerID="f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6" Oct 11 03:51:59 crc kubenswrapper[4754]: E1011 03:51:59.743115 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6\": container with ID starting with f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6 not found: ID does not exist" containerID="f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6" Oct 11 03:51:59 crc kubenswrapper[4754]: I1011 03:51:59.743197 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6"} err="failed to get container status \"f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6\": rpc error: code = NotFound desc = could not find container \"f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6\": container with ID starting with f03055efbc348094dc69fa4ad251bd2087694453e6032fba39c33aae07c171a6 not found: ID does not exist" Oct 11 03:52:01 crc kubenswrapper[4754]: I1011 03:52:01.096855 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" path="/var/lib/kubelet/pods/f6151612-e467-4803-ac1c-f10a3b7dc9b9/volumes" Oct 11 03:52:15 crc kubenswrapper[4754]: I1011 03:52:15.793490 4754 generic.go:334] "Generic (PLEG): container finished" podID="347dff33-1661-4ed0-97b0-4248e8648f6a" containerID="4c2f54d27095a3e964c95c324e2255ee004bef1dea5493d5e71af2ba7be3410f" exitCode=0 Oct 11 03:52:15 crc kubenswrapper[4754]: I1011 03:52:15.794089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" event={"ID":"347dff33-1661-4ed0-97b0-4248e8648f6a","Type":"ContainerDied","Data":"4c2f54d27095a3e964c95c324e2255ee004bef1dea5493d5e71af2ba7be3410f"} Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.264995 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405174 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4pk5\" (UniqueName: \"kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405255 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405419 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405534 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.405580 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph\") pod \"347dff33-1661-4ed0-97b0-4248e8648f6a\" (UID: \"347dff33-1661-4ed0-97b0-4248e8648f6a\") " Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.411183 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.411206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph" (OuterVolumeSpecName: "ceph") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.412165 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5" (OuterVolumeSpecName: "kube-api-access-l4pk5") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "kube-api-access-l4pk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.431381 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.436040 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory" (OuterVolumeSpecName: "inventory") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.437461 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "347dff33-1661-4ed0-97b0-4248e8648f6a" (UID: "347dff33-1661-4ed0-97b0-4248e8648f6a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507607 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507644 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507656 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507667 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4pk5\" (UniqueName: \"kubernetes.io/projected/347dff33-1661-4ed0-97b0-4248e8648f6a-kube-api-access-l4pk5\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507681 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.507695 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/347dff33-1661-4ed0-97b0-4248e8648f6a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.813200 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" event={"ID":"347dff33-1661-4ed0-97b0-4248e8648f6a","Type":"ContainerDied","Data":"bb54b6761ec166e0966ecc22aca3620b289772ef5c967d41f3356879ab62ecf6"} Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.813533 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb54b6761ec166e0966ecc22aca3620b289772ef5c967d41f3356879ab62ecf6" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.813253 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b59cc" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.922385 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb"] Oct 11 03:52:17 crc kubenswrapper[4754]: E1011 03:52:17.922763 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="extract-utilities" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.922779 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="extract-utilities" Oct 11 03:52:17 crc kubenswrapper[4754]: E1011 03:52:17.922799 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="registry-server" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.922806 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="registry-server" Oct 11 03:52:17 crc kubenswrapper[4754]: E1011 03:52:17.922825 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="347dff33-1661-4ed0-97b0-4248e8648f6a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.922833 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="347dff33-1661-4ed0-97b0-4248e8648f6a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 11 03:52:17 crc kubenswrapper[4754]: E1011 03:52:17.922847 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="extract-content" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.922853 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="extract-content" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.923072 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="347dff33-1661-4ed0-97b0-4248e8648f6a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.923090 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6151612-e467-4803-ac1c-f10a3b7dc9b9" containerName="registry-server" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.923831 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.928033 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.928339 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.928406 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.934434 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.934466 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wf68j" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.934516 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.934439 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.935244 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.935473 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:52:17 crc kubenswrapper[4754]: I1011 03:52:17.943746 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb"] Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117613 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117683 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117747 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117822 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.117910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psbss\" (UniqueName: \"kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.118064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.118098 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.118165 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.118219 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.118404 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.219846 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.219904 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.219934 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.219975 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.219998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220020 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220121 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220151 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220168 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.220191 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psbss\" (UniqueName: \"kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.221614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.223658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.225875 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.226145 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.226951 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.233576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.233651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.233918 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.234564 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.234735 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.237946 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psbss\" (UniqueName: \"kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.241955 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.767949 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb"] Oct 11 03:52:18 crc kubenswrapper[4754]: I1011 03:52:18.821763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" event={"ID":"ff20e060-f35b-4a0f-8861-a7befb21032e","Type":"ContainerStarted","Data":"2542556efc1f8b4152be368bc0819f983f10cf1425f6ff03cb9017b173f9e93b"} Oct 11 03:52:19 crc kubenswrapper[4754]: I1011 03:52:19.836031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" event={"ID":"ff20e060-f35b-4a0f-8861-a7befb21032e","Type":"ContainerStarted","Data":"b9ea69961d16e6a253af4f8ff320762fd9acf296d7b3440f91dea2acd52599f8"} Oct 11 03:52:19 crc kubenswrapper[4754]: I1011 03:52:19.899446 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" podStartSLOduration=2.435892665 podStartE2EDuration="2.899406097s" podCreationTimestamp="2025-10-11 03:52:17 +0000 UTC" firstStartedPulling="2025-10-11 03:52:18.778952859 +0000 UTC m=+2786.337897644" lastFinishedPulling="2025-10-11 03:52:19.242466291 +0000 UTC m=+2786.801411076" observedRunningTime="2025-10-11 03:52:19.871696996 +0000 UTC m=+2787.430641791" watchObservedRunningTime="2025-10-11 03:52:19.899406097 +0000 UTC m=+2787.458350892" Oct 11 03:53:00 crc kubenswrapper[4754]: I1011 03:53:00.736330 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:53:00 crc kubenswrapper[4754]: I1011 03:53:00.736917 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:53:30 crc kubenswrapper[4754]: I1011 03:53:30.736907 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:53:30 crc kubenswrapper[4754]: I1011 03:53:30.737670 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.736948 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.737911 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.738009 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.739305 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.739383 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4" gracePeriod=600 Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.883750 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4" exitCode=0 Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.883805 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4"} Oct 11 03:54:00 crc kubenswrapper[4754]: I1011 03:54:00.883843 4754 scope.go:117] "RemoveContainer" containerID="56b31384028062c1a11fd6294d572bcbc52b342d542a53198b94dd35e631b5f6" Oct 11 03:54:01 crc kubenswrapper[4754]: I1011 03:54:01.896831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1"} Oct 11 03:56:09 crc kubenswrapper[4754]: I1011 03:56:09.245552 4754 generic.go:334] "Generic (PLEG): container finished" podID="ff20e060-f35b-4a0f-8861-a7befb21032e" containerID="b9ea69961d16e6a253af4f8ff320762fd9acf296d7b3440f91dea2acd52599f8" exitCode=0 Oct 11 03:56:09 crc kubenswrapper[4754]: I1011 03:56:09.245624 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" event={"ID":"ff20e060-f35b-4a0f-8861-a7befb21032e","Type":"ContainerDied","Data":"b9ea69961d16e6a253af4f8ff320762fd9acf296d7b3440f91dea2acd52599f8"} Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.700126 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828304 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828393 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828448 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828484 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psbss\" (UniqueName: \"kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828528 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828617 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828651 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828773 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.828840 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph\") pod \"ff20e060-f35b-4a0f-8861-a7befb21032e\" (UID: \"ff20e060-f35b-4a0f-8861-a7befb21032e\") " Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.837769 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.849116 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss" (OuterVolumeSpecName: "kube-api-access-psbss") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "kube-api-access-psbss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.851826 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph" (OuterVolumeSpecName: "ceph") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.857635 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.862320 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory" (OuterVolumeSpecName: "inventory") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.866526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.867705 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.868818 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.872321 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.880424 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.894836 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ff20e060-f35b-4a0f-8861-a7befb21032e" (UID: "ff20e060-f35b-4a0f-8861-a7befb21032e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934288 4754 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934372 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-inventory\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934389 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934415 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934430 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934444 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934458 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934473 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934487 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psbss\" (UniqueName: \"kubernetes.io/projected/ff20e060-f35b-4a0f-8861-a7befb21032e-kube-api-access-psbss\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934499 4754 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ff20e060-f35b-4a0f-8861-a7befb21032e-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:10 crc kubenswrapper[4754]: I1011 03:56:10.934513 4754 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff20e060-f35b-4a0f-8861-a7befb21032e-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:11 crc kubenswrapper[4754]: I1011 03:56:11.280505 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" event={"ID":"ff20e060-f35b-4a0f-8861-a7befb21032e","Type":"ContainerDied","Data":"2542556efc1f8b4152be368bc0819f983f10cf1425f6ff03cb9017b173f9e93b"} Oct 11 03:56:11 crc kubenswrapper[4754]: I1011 03:56:11.280868 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2542556efc1f8b4152be368bc0819f983f10cf1425f6ff03cb9017b173f9e93b" Oct 11 03:56:11 crc kubenswrapper[4754]: I1011 03:56:11.280622 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.044909 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 11 03:56:26 crc kubenswrapper[4754]: E1011 03:56:26.045897 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff20e060-f35b-4a0f-8861-a7befb21032e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.045913 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff20e060-f35b-4a0f-8861-a7befb21032e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.046111 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff20e060-f35b-4a0f-8861-a7befb21032e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.047118 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.050490 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.051260 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.079915 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201125 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201210 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201236 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201700 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201804 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.201877 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202000 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202081 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202131 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202161 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-run\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5tn6\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-kube-api-access-j5tn6\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202360 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.202390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.221037 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.222587 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.225773 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.243911 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305422 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305611 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-dev\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305677 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305699 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305726 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305770 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-scripts\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305841 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.305843 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306015 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306086 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306128 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-run\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306174 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306190 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-ceph\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-run\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306435 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-run\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306495 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5tn6\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-kube-api-access-j5tn6\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306676 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-lib-modules\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306850 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.306946 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data-custom\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307012 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-nvme\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307090 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307114 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-sys\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307155 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307229 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307278 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t949g\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-kube-api-access-t949g\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307328 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307337 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307521 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ae6da214-d13e-4fde-9e99-1e8db2f6e248-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.307776 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.314406 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.314812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.327772 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.327898 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.328420 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae6da214-d13e-4fde-9e99-1e8db2f6e248-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.331271 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5tn6\" (UniqueName: \"kubernetes.io/projected/ae6da214-d13e-4fde-9e99-1e8db2f6e248-kube-api-access-j5tn6\") pod \"cinder-volume-volume1-0\" (UID: \"ae6da214-d13e-4fde-9e99-1e8db2f6e248\") " pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.372095 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410564 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-run\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410653 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-lib-modules\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410718 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410714 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-run\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410761 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data-custom\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410792 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410758 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-lib-modules\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410905 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.410815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411031 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-nvme\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-sys\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411105 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-nvme\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411138 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411193 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411221 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-sys\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411233 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t949g\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-kube-api-access-t949g\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411268 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411366 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-dev\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411664 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-scripts\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-dev\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411792 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61b305b9-2dcf-4961-a33d-d93a242b01cf-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.411820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-ceph\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.417181 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-scripts\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.418655 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data-custom\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.421210 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-ceph\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.421509 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.426676 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b305b9-2dcf-4961-a33d-d93a242b01cf-config-data\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.428296 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t949g\" (UniqueName: \"kubernetes.io/projected/61b305b9-2dcf-4961-a33d-d93a242b01cf-kube-api-access-t949g\") pod \"cinder-backup-0\" (UID: \"61b305b9-2dcf-4961-a33d-d93a242b01cf\") " pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.542560 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.700196 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-t9ct9"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.701904 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.723408 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t9ct9"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.812742 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.814590 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.824288 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.824332 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-b4xcv" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.824632 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.824756 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.825701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zgtr\" (UniqueName: \"kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr\") pod \"manila-db-create-t9ct9\" (UID: \"10b2f10c-b480-4250-b44c-883892870d1b\") " pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.929771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.929870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.929945 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zgtr\" (UniqueName: \"kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr\") pod \"manila-db-create-t9ct9\" (UID: \"10b2f10c-b480-4250-b44c-883892870d1b\") " pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.941634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.942014 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.942308 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8jtl\" (UniqueName: \"kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.942361 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.962083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.975186 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.975597 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.975791 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ntk29" Oct 11 03:56:26 crc kubenswrapper[4754]: I1011 03:56:26.975945 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.019494 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.035176 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zgtr\" (UniqueName: \"kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr\") pod \"manila-db-create-t9ct9\" (UID: \"10b2f10c-b480-4250-b44c-883892870d1b\") " pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.044940 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046214 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046247 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046284 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046322 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk5ff\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046411 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046429 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046474 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046513 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.046570 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8jtl\" (UniqueName: \"kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.047927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.048575 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.049857 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.054021 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.058563 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.066187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8jtl\" (UniqueName: \"kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl\") pod \"horizon-86df4d8c9c-qtbx4\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.066269 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.068834 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.078441 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.080451 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.088018 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.088635 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.107171 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.124531 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:56:27 crc kubenswrapper[4754]: W1011 03:56:27.138780 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae6da214_d13e_4fde_9e99_1e8db2f6e248.slice/crio-370edf15d809d4d7449bbbfc5d398a9fc4942037d707f79431bba2b5dbf29bcc WatchSource:0}: Error finding container 370edf15d809d4d7449bbbfc5d398a9fc4942037d707f79431bba2b5dbf29bcc: Status 404 returned error can't find the container with id 370edf15d809d4d7449bbbfc5d398a9fc4942037d707f79431bba2b5dbf29bcc Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.151543 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.151749 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152317 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152352 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152407 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8lpt\" (UniqueName: \"kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152457 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152484 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152510 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152616 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152662 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152724 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152791 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152807 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbcf\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152843 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.152954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.153042 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.153093 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.153126 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.154118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.154186 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.154195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.155038 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.156812 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.156895 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk5ff\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.159341 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.160261 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.161981 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.162313 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.163225 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.164685 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.180974 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.193027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk5ff\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.234393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260434 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8lpt\" (UniqueName: \"kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260481 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260522 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260600 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260694 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbcf\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260720 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260760 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.260978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.261614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.261948 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.263159 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.263719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.264109 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.264251 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.269846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.270643 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.272769 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.273586 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.291367 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8lpt\" (UniqueName: \"kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt\") pod \"horizon-785c858b67-vvjnn\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.292492 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.294145 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.304735 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbcf\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.335600 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.345903 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.396664 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.437623 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.445842 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.544163 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"ae6da214-d13e-4fde-9e99-1e8db2f6e248","Type":"ContainerStarted","Data":"370edf15d809d4d7449bbbfc5d398a9fc4942037d707f79431bba2b5dbf29bcc"} Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.596918 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t9ct9"] Oct 11 03:56:27 crc kubenswrapper[4754]: W1011 03:56:27.605992 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10b2f10c_b480_4250_b44c_883892870d1b.slice/crio-da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb WatchSource:0}: Error finding container da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb: Status 404 returned error can't find the container with id da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.795238 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:56:27 crc kubenswrapper[4754]: W1011 03:56:27.931055 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c190463_553c_41d4_ac6c_122832a2b286.slice/crio-fa4384dff4d5eac69ad3c4833fb8d0fd3fc3e3a03f0b1745c369c1674229ac90 WatchSource:0}: Error finding container fa4384dff4d5eac69ad3c4833fb8d0fd3fc3e3a03f0b1745c369c1674229ac90: Status 404 returned error can't find the container with id fa4384dff4d5eac69ad3c4833fb8d0fd3fc3e3a03f0b1745c369c1674229ac90 Oct 11 03:56:27 crc kubenswrapper[4754]: I1011 03:56:27.936288 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.265692 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:28 crc kubenswrapper[4754]: W1011 03:56:28.282089 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b99fbae_638f_4abe_a92a_e1fa01770365.slice/crio-9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b WatchSource:0}: Error finding container 9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b: Status 404 returned error can't find the container with id 9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.573408 4754 generic.go:334] "Generic (PLEG): container finished" podID="10b2f10c-b480-4250-b44c-883892870d1b" containerID="538393f387053279ccb6b2298bda16aec0b28fb06233a7441bedfdc9f57a6fdf" exitCode=0 Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.573500 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t9ct9" event={"ID":"10b2f10c-b480-4250-b44c-883892870d1b","Type":"ContainerDied","Data":"538393f387053279ccb6b2298bda16aec0b28fb06233a7441bedfdc9f57a6fdf"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.574366 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t9ct9" event={"ID":"10b2f10c-b480-4250-b44c-883892870d1b","Type":"ContainerStarted","Data":"da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.584877 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"ae6da214-d13e-4fde-9e99-1e8db2f6e248","Type":"ContainerStarted","Data":"a582bf16286cb86a8df63cf8c195067265eed1b42e10a9e5d2f1d736d0551b26"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.588796 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"61b305b9-2dcf-4961-a33d-d93a242b01cf","Type":"ContainerStarted","Data":"f2e5cbfb0686b65695fefa6841596004b424c8d9534f76005fac5400fbba1b9e"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.590656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerStarted","Data":"fa4384dff4d5eac69ad3c4833fb8d0fd3fc3e3a03f0b1745c369c1674229ac90"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.599808 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerStarted","Data":"9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b"} Oct 11 03:56:28 crc kubenswrapper[4754]: I1011 03:56:28.601180 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerStarted","Data":"d535bdeb161598f52260227f909eff5e52776c7a7268d00df21db5a2b3fadf5c"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.154549 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.624638 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"61b305b9-2dcf-4961-a33d-d93a242b01cf","Type":"ContainerStarted","Data":"41eb6dc37d363179df0835bd1225ee2d7005adcef5e23700a819a69a5f341003"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.625340 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"61b305b9-2dcf-4961-a33d-d93a242b01cf","Type":"ContainerStarted","Data":"f45510f527e22a2bb8656a00b6c8e80e0d65832e1f114bbe17de943ba046371e"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.657302 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerStarted","Data":"ef182d165e1dfb0b6a4e3aee7b3abf44e5296391b218730ea0682f96042571d1"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.658775 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.695434506 podStartE2EDuration="3.658754175s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="2025-10-11 03:56:27.56217152 +0000 UTC m=+3035.121116305" lastFinishedPulling="2025-10-11 03:56:28.525491189 +0000 UTC m=+3036.084435974" observedRunningTime="2025-10-11 03:56:29.654461444 +0000 UTC m=+3037.213406229" watchObservedRunningTime="2025-10-11 03:56:29.658754175 +0000 UTC m=+3037.217698960" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.665600 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerStarted","Data":"f02ad7f0be515f242c2e6a68855c00dfd21770773115f4e650a475a55312480d"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.691388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"ae6da214-d13e-4fde-9e99-1e8db2f6e248","Type":"ContainerStarted","Data":"f019a099b6b6f0f184671cdf2a5dff2b9432e3f709964eb2be1f6c9f7df94e7f"} Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.759659 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.792520 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.811175491 podStartE2EDuration="3.792492596s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="2025-10-11 03:56:27.151183753 +0000 UTC m=+3034.710128528" lastFinishedPulling="2025-10-11 03:56:28.132500848 +0000 UTC m=+3035.691445633" observedRunningTime="2025-10-11 03:56:29.760821135 +0000 UTC m=+3037.319765930" watchObservedRunningTime="2025-10-11 03:56:29.792492596 +0000 UTC m=+3037.351437381" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.811397 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.813296 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.819340 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.852065 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.873415 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940254 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940289 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjpbq\" (UniqueName: \"kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940422 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940624 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.940766 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.950329 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.959211 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8655c657c6-4mw6n"] Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.963352 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:29 crc kubenswrapper[4754]: I1011 03:56:29.981032 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8655c657c6-4mw6n"] Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.043304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-logs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.043575 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.043696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvtv\" (UniqueName: \"kubernetes.io/projected/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-kube-api-access-wsvtv\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.043778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-scripts\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-combined-ca-bundle\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-secret-key\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044457 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044483 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.044908 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.045059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-tls-certs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.045089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.045226 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjpbq\" (UniqueName: \"kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.045318 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-config-data\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.046365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.047745 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.059570 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.064049 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.067812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.073829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjpbq\" (UniqueName: \"kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq\") pod \"horizon-586ffbcfb8-rqrkz\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.141909 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.152336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvtv\" (UniqueName: \"kubernetes.io/projected/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-kube-api-access-wsvtv\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.152406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-scripts\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.152470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-combined-ca-bundle\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.152572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-secret-key\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.152810 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-tls-certs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.153024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-config-data\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.153070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-logs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.153635 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-logs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.157755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-combined-ca-bundle\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.157918 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-scripts\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.158925 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-config-data\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.162524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-secret-key\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.166673 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-horizon-tls-certs\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.191802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvtv\" (UniqueName: \"kubernetes.io/projected/96d0a63b-15e8-4830-8fd6-1bd47e6ed973-kube-api-access-wsvtv\") pod \"horizon-8655c657c6-4mw6n\" (UID: \"96d0a63b-15e8-4830-8fd6-1bd47e6ed973\") " pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.268765 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.316865 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.359667 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zgtr\" (UniqueName: \"kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr\") pod \"10b2f10c-b480-4250-b44c-883892870d1b\" (UID: \"10b2f10c-b480-4250-b44c-883892870d1b\") " Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.371477 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr" (OuterVolumeSpecName: "kube-api-access-4zgtr") pod "10b2f10c-b480-4250-b44c-883892870d1b" (UID: "10b2f10c-b480-4250-b44c-883892870d1b"). InnerVolumeSpecName "kube-api-access-4zgtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.465797 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zgtr\" (UniqueName: \"kubernetes.io/projected/10b2f10c-b480-4250-b44c-883892870d1b-kube-api-access-4zgtr\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.799996 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.800374 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.813035 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t9ct9" event={"ID":"10b2f10c-b480-4250-b44c-883892870d1b","Type":"ContainerDied","Data":"da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb"} Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.813102 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da12a889502f8c3092e23a6dcd8fcaa87c5eb04aaa2fe27dd155fd2cce9303bb" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.813240 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t9ct9" Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.842317 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerStarted","Data":"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114"} Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.855632 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.893368 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerStarted","Data":"6181c8f3937f5a6c3d5e8285ccd45c91d5273e5df3c8cd3eee89a4de7035448f"} Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.896485 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-log" containerID="cri-o://f02ad7f0be515f242c2e6a68855c00dfd21770773115f4e650a475a55312480d" gracePeriod=30 Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.897225 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-httpd" containerID="cri-o://6181c8f3937f5a6c3d5e8285ccd45c91d5273e5df3c8cd3eee89a4de7035448f" gracePeriod=30 Oct 11 03:56:30 crc kubenswrapper[4754]: I1011 03:56:30.955891 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.95586757 podStartE2EDuration="4.95586757s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:56:30.950629933 +0000 UTC m=+3038.509574718" watchObservedRunningTime="2025-10-11 03:56:30.95586757 +0000 UTC m=+3038.514812345" Oct 11 03:56:31 crc kubenswrapper[4754]: W1011 03:56:31.147341 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96d0a63b_15e8_4830_8fd6_1bd47e6ed973.slice/crio-139acbc5e28c2610b18a77ffed1c295e3fcf103e8d07897d4840eb2cef6f9401 WatchSource:0}: Error finding container 139acbc5e28c2610b18a77ffed1c295e3fcf103e8d07897d4840eb2cef6f9401: Status 404 returned error can't find the container with id 139acbc5e28c2610b18a77ffed1c295e3fcf103e8d07897d4840eb2cef6f9401 Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.169490 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8655c657c6-4mw6n"] Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.375522 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.543160 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905183 4754 generic.go:334] "Generic (PLEG): container finished" podID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerID="6181c8f3937f5a6c3d5e8285ccd45c91d5273e5df3c8cd3eee89a4de7035448f" exitCode=0 Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905620 4754 generic.go:334] "Generic (PLEG): container finished" podID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerID="f02ad7f0be515f242c2e6a68855c00dfd21770773115f4e650a475a55312480d" exitCode=143 Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905660 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerDied","Data":"6181c8f3937f5a6c3d5e8285ccd45c91d5273e5df3c8cd3eee89a4de7035448f"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905698 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerDied","Data":"f02ad7f0be515f242c2e6a68855c00dfd21770773115f4e650a475a55312480d"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905711 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b99fbae-638f-4abe-a92a-e1fa01770365","Type":"ContainerDied","Data":"9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.905720 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a7551d31faa546ccc2c043ab5eed60f67ac1e7663cf91f7522201a0f83d3e8b" Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.907289 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8655c657c6-4mw6n" event={"ID":"96d0a63b-15e8-4830-8fd6-1bd47e6ed973","Type":"ContainerStarted","Data":"139acbc5e28c2610b18a77ffed1c295e3fcf103e8d07897d4840eb2cef6f9401"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.909226 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerStarted","Data":"2ab2a5ab20289474552d2b3f07531fb8edc4b405b74e44cb24944fe5bc4132a6"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.914855 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-log" containerID="cri-o://8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" gracePeriod=30 Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.914908 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-httpd" containerID="cri-o://5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" gracePeriod=30 Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.915403 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerStarted","Data":"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f"} Oct 11 03:56:31 crc kubenswrapper[4754]: I1011 03:56:31.955528 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.95550477 podStartE2EDuration="5.95550477s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:56:31.950158479 +0000 UTC m=+3039.509103274" watchObservedRunningTime="2025-10-11 03:56:31.95550477 +0000 UTC m=+3039.514449555" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.006640 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.158086 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.158420 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159233 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159392 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk5ff\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159529 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159567 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159611 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.159733 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run\") pod \"9b99fbae-638f-4abe-a92a-e1fa01770365\" (UID: \"9b99fbae-638f-4abe-a92a-e1fa01770365\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.160729 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.162173 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs" (OuterVolumeSpecName: "logs") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.167093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff" (OuterVolumeSpecName: "kube-api-access-pk5ff") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "kube-api-access-pk5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.167761 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph" (OuterVolumeSpecName: "ceph") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.167856 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.167991 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts" (OuterVolumeSpecName: "scripts") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.204214 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.220884 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.221725 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data" (OuterVolumeSpecName: "config-data") pod "9b99fbae-638f-4abe-a92a-e1fa01770365" (UID: "9b99fbae-638f-4abe-a92a-e1fa01770365"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263662 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263710 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263757 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263770 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263781 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk5ff\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-kube-api-access-pk5ff\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263792 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b99fbae-638f-4abe-a92a-e1fa01770365-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263801 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263813 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b99fbae-638f-4abe-a92a-e1fa01770365-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.263824 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b99fbae-638f-4abe-a92a-e1fa01770365-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.289534 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.368636 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.697851 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.777901 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778067 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778155 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778215 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778285 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778334 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778421 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.778813 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbcf\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf\") pod \"0c67f627-2cb6-4835-bae9-3ca35320eb75\" (UID: \"0c67f627-2cb6-4835-bae9-3ca35320eb75\") " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.779311 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs" (OuterVolumeSpecName: "logs") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.779379 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.781244 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.781275 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c67f627-2cb6-4835-bae9-3ca35320eb75-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.785093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph" (OuterVolumeSpecName: "ceph") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.785884 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts" (OuterVolumeSpecName: "scripts") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.787405 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf" (OuterVolumeSpecName: "kube-api-access-vzbcf") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "kube-api-access-vzbcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.800228 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.815050 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.867539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data" (OuterVolumeSpecName: "config-data") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.873850 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0c67f627-2cb6-4835-bae9-3ca35320eb75" (UID: "0c67f627-2cb6-4835-bae9-3ca35320eb75"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884632 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884696 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884714 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884727 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884737 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbcf\" (UniqueName: \"kubernetes.io/projected/0c67f627-2cb6-4835-bae9-3ca35320eb75-kube-api-access-vzbcf\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884748 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c67f627-2cb6-4835-bae9-3ca35320eb75-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.884807 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.915895 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.930822 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerID="5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" exitCode=0 Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.930868 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerID="8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" exitCode=143 Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.930932 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.931367 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.934165 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerDied","Data":"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f"} Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.934244 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerDied","Data":"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114"} Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.934260 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c67f627-2cb6-4835-bae9-3ca35320eb75","Type":"ContainerDied","Data":"ef182d165e1dfb0b6a4e3aee7b3abf44e5296391b218730ea0682f96042571d1"} Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.934286 4754 scope.go:117] "RemoveContainer" containerID="5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" Oct 11 03:56:32 crc kubenswrapper[4754]: I1011 03:56:32.987341 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:32.998714 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:32.998859 4754 scope.go:117] "RemoveContainer" containerID="8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.009879 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.022231 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.035885 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.044820 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.045728 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.045763 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.045780 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.045789 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.045823 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.045832 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.045868 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10b2f10c-b480-4250-b44c-883892870d1b" containerName="mariadb-database-create" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.045878 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="10b2f10c-b480-4250-b44c-883892870d1b" containerName="mariadb-database-create" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.045895 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.045905 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.046213 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.046250 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.046264 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="10b2f10c-b480-4250-b44c-883892870d1b" containerName="mariadb-database-create" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.046283 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-httpd" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.046303 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" containerName="glance-log" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.047938 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.052731 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.058261 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.059090 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.059668 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.061242 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.062415 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.062667 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.062849 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ntk29" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.064370 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.072624 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.114844 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c67f627-2cb6-4835-bae9-3ca35320eb75" path="/var/lib/kubelet/pods/0c67f627-2cb6-4835-bae9-3ca35320eb75/volumes" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.118379 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b99fbae-638f-4abe-a92a-e1fa01770365" path="/var/lib/kubelet/pods/9b99fbae-638f-4abe-a92a-e1fa01770365/volumes" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.122288 4754 scope.go:117] "RemoveContainer" containerID="5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.134571 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f\": container with ID starting with 5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f not found: ID does not exist" containerID="5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.134650 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f"} err="failed to get container status \"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f\": rpc error: code = NotFound desc = could not find container \"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f\": container with ID starting with 5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f not found: ID does not exist" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.134822 4754 scope.go:117] "RemoveContainer" containerID="8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" Oct 11 03:56:33 crc kubenswrapper[4754]: E1011 03:56:33.138781 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114\": container with ID starting with 8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114 not found: ID does not exist" containerID="8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.138862 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114"} err="failed to get container status \"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114\": rpc error: code = NotFound desc = could not find container \"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114\": container with ID starting with 8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114 not found: ID does not exist" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.138892 4754 scope.go:117] "RemoveContainer" containerID="5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.139643 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f"} err="failed to get container status \"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f\": rpc error: code = NotFound desc = could not find container \"5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f\": container with ID starting with 5fdd9e7636dd3f65e6d3abdbc96ef518955ae62941b55dd81d1e54a8ce80f52f not found: ID does not exist" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.139815 4754 scope.go:117] "RemoveContainer" containerID="8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.140182 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114"} err="failed to get container status \"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114\": rpc error: code = NotFound desc = could not find container \"8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114\": container with ID starting with 8b9fd331cfb9a2bbb948af1b3a78dfa2f828859f8358f199857a4654678de114 not found: ID does not exist" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.193150 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-ceph\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.194078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vb4w\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-kube-api-access-8vb4w\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.197644 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.198812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.198888 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199165 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199359 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199427 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199512 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k559h\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-kube-api-access-k559h\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-logs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199668 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199744 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.199820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.200002 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.200067 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.302837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.302882 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.302919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.302938 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.302986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k559h\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-kube-api-access-k559h\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303083 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-logs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303198 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303235 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303255 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-ceph\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303273 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vb4w\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-kube-api-access-8vb4w\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303324 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.308281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.312079 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.312391 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.312705 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.313036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-logs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.303332 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.313624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.313794 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.313867 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.314464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8cc0a22-3499-4f27-9250-70c5a233a95f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.314463 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.315169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.315350 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.317348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-ceph\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.322213 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.327870 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k559h\" (UniqueName: \"kubernetes.io/projected/c8cc0a22-3499-4f27-9250-70c5a233a95f-kube-api-access-k559h\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.328168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cc0a22-3499-4f27-9250-70c5a233a95f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.331823 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vb4w\" (UniqueName: \"kubernetes.io/projected/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-kube-api-access-8vb4w\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.346118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.359821 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca\") " pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.373604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8cc0a22-3499-4f27-9250-70c5a233a95f\") " pod="openstack/glance-default-internal-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.406808 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 03:56:33 crc kubenswrapper[4754]: I1011 03:56:33.420421 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:34 crc kubenswrapper[4754]: I1011 03:56:34.096932 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 03:56:34 crc kubenswrapper[4754]: W1011 03:56:34.102521 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8cc0a22_3499_4f27_9250_70c5a233a95f.slice/crio-91f337890548122b72ab1cdefb5416e380c26164cafc1974978c9b94f8f5caa7 WatchSource:0}: Error finding container 91f337890548122b72ab1cdefb5416e380c26164cafc1974978c9b94f8f5caa7: Status 404 returned error can't find the container with id 91f337890548122b72ab1cdefb5416e380c26164cafc1974978c9b94f8f5caa7 Oct 11 03:56:34 crc kubenswrapper[4754]: I1011 03:56:34.224028 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 03:56:34 crc kubenswrapper[4754]: I1011 03:56:34.955546 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8cc0a22-3499-4f27-9250-70c5a233a95f","Type":"ContainerStarted","Data":"91f337890548122b72ab1cdefb5416e380c26164cafc1974978c9b94f8f5caa7"} Oct 11 03:56:34 crc kubenswrapper[4754]: I1011 03:56:34.962562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca","Type":"ContainerStarted","Data":"97a8b0143b2c6f042be92ca87371d80daa441e363933ce87bff5889313b2cbf8"} Oct 11 03:56:35 crc kubenswrapper[4754]: I1011 03:56:35.976023 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca","Type":"ContainerStarted","Data":"e11a1d9b832e55566a64f5cca27edf0d792237fcf140ecbadbb52c08b2e27c95"} Oct 11 03:56:35 crc kubenswrapper[4754]: I1011 03:56:35.980323 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8cc0a22-3499-4f27-9250-70c5a233a95f","Type":"ContainerStarted","Data":"e7a2f30f4a18efc02ccf47b98699ac9b94af59ad7228162e70f7b5ac37eabadd"} Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.607571 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.816602 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.851078 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-ed00-account-create-q2rgc"] Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.852359 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.859638 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.867271 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ed00-account-create-q2rgc"] Oct 11 03:56:36 crc kubenswrapper[4754]: I1011 03:56:36.906825 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w8fv\" (UniqueName: \"kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv\") pod \"manila-ed00-account-create-q2rgc\" (UID: \"f7055e1d-ffb5-4cca-9a53-2400675059a0\") " pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:37 crc kubenswrapper[4754]: I1011 03:56:37.010039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w8fv\" (UniqueName: \"kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv\") pod \"manila-ed00-account-create-q2rgc\" (UID: \"f7055e1d-ffb5-4cca-9a53-2400675059a0\") " pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:37 crc kubenswrapper[4754]: I1011 03:56:37.032856 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w8fv\" (UniqueName: \"kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv\") pod \"manila-ed00-account-create-q2rgc\" (UID: \"f7055e1d-ffb5-4cca-9a53-2400675059a0\") " pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:37 crc kubenswrapper[4754]: I1011 03:56:37.174142 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:40 crc kubenswrapper[4754]: I1011 03:56:40.693200 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ed00-account-create-q2rgc"] Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.040743 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerStarted","Data":"88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.043664 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8655c657c6-4mw6n" event={"ID":"96d0a63b-15e8-4830-8fd6-1bd47e6ed973","Type":"ContainerStarted","Data":"b81978f63421cae3b3fc9cb3edbd91051944e8f6090cada133b8b08f19a220d4"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.043728 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8655c657c6-4mw6n" event={"ID":"96d0a63b-15e8-4830-8fd6-1bd47e6ed973","Type":"ContainerStarted","Data":"56f84504de073f7727bebb75f10e520cc8155235e66cd1dfa71bfa985b00e5e4"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.050612 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7055e1d-ffb5-4cca-9a53-2400675059a0" containerID="d9b934ea544b50dbcffeba14c6bbe95f45b7f0c3562e7cb45b752331ea5e7da7" exitCode=0 Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.050712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ed00-account-create-q2rgc" event={"ID":"f7055e1d-ffb5-4cca-9a53-2400675059a0","Type":"ContainerDied","Data":"d9b934ea544b50dbcffeba14c6bbe95f45b7f0c3562e7cb45b752331ea5e7da7"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.050751 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ed00-account-create-q2rgc" event={"ID":"f7055e1d-ffb5-4cca-9a53-2400675059a0","Type":"ContainerStarted","Data":"06be9f75840c36c64b84cebba80c3698c31f45bbaa1def5e09a36bd6a90b3c7c"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.054280 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerStarted","Data":"ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.062120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerStarted","Data":"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121"} Oct 11 03:56:41 crc kubenswrapper[4754]: I1011 03:56:41.080645 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8655c657c6-4mw6n" podStartSLOduration=2.886196871 podStartE2EDuration="12.080617095s" podCreationTimestamp="2025-10-11 03:56:29 +0000 UTC" firstStartedPulling="2025-10-11 03:56:31.151590494 +0000 UTC m=+3038.710535279" lastFinishedPulling="2025-10-11 03:56:40.346010708 +0000 UTC m=+3047.904955503" observedRunningTime="2025-10-11 03:56:41.078580337 +0000 UTC m=+3048.637525122" watchObservedRunningTime="2025-10-11 03:56:41.080617095 +0000 UTC m=+3048.639561880" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.085093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerStarted","Data":"3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13"} Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.085196 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86df4d8c9c-qtbx4" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon-log" containerID="cri-o://ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667" gracePeriod=30 Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.085262 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86df4d8c9c-qtbx4" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon" containerID="cri-o://3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13" gracePeriod=30 Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.091433 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca","Type":"ContainerStarted","Data":"e07186ac13d7eaf3807ebff67aa65364d17f5a526e61c8cefa040b79d9ef6f6e"} Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.095454 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerStarted","Data":"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e"} Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.099781 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8cc0a22-3499-4f27-9250-70c5a233a95f","Type":"ContainerStarted","Data":"fed3540a6f00bc2b48b8c07d99655c4f7f7a0b32a96f076d5d7688561d550594"} Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.103871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerStarted","Data":"8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07"} Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.103989 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-785c858b67-vvjnn" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon-log" containerID="cri-o://88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe" gracePeriod=30 Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.104008 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-785c858b67-vvjnn" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon" containerID="cri-o://8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07" gracePeriod=30 Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.115900 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-86df4d8c9c-qtbx4" podStartSLOduration=3.471830769 podStartE2EDuration="16.115876036s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="2025-10-11 03:56:27.804169785 +0000 UTC m=+3035.363114570" lastFinishedPulling="2025-10-11 03:56:40.448215052 +0000 UTC m=+3048.007159837" observedRunningTime="2025-10-11 03:56:42.114796796 +0000 UTC m=+3049.673741591" watchObservedRunningTime="2025-10-11 03:56:42.115876036 +0000 UTC m=+3049.674820821" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.165006 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.164954456 podStartE2EDuration="10.164954456s" podCreationTimestamp="2025-10-11 03:56:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:56:42.151054766 +0000 UTC m=+3049.709999561" watchObservedRunningTime="2025-10-11 03:56:42.164954456 +0000 UTC m=+3049.723899251" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.187000 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.186955785 podStartE2EDuration="9.186955785s" podCreationTimestamp="2025-10-11 03:56:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:56:42.178327802 +0000 UTC m=+3049.737272587" watchObservedRunningTime="2025-10-11 03:56:42.186955785 +0000 UTC m=+3049.745900570" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.210274 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-785c858b67-vvjnn" podStartSLOduration=3.845522148 podStartE2EDuration="16.21024293s" podCreationTimestamp="2025-10-11 03:56:26 +0000 UTC" firstStartedPulling="2025-10-11 03:56:27.933567554 +0000 UTC m=+3035.492512339" lastFinishedPulling="2025-10-11 03:56:40.298288336 +0000 UTC m=+3047.857233121" observedRunningTime="2025-10-11 03:56:42.203893171 +0000 UTC m=+3049.762837956" watchObservedRunningTime="2025-10-11 03:56:42.21024293 +0000 UTC m=+3049.769187735" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.234652 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-586ffbcfb8-rqrkz" podStartSLOduration=3.802341953 podStartE2EDuration="13.234633236s" podCreationTimestamp="2025-10-11 03:56:29 +0000 UTC" firstStartedPulling="2025-10-11 03:56:30.913803167 +0000 UTC m=+3038.472747952" lastFinishedPulling="2025-10-11 03:56:40.34609445 +0000 UTC m=+3047.905039235" observedRunningTime="2025-10-11 03:56:42.229939444 +0000 UTC m=+3049.788884229" watchObservedRunningTime="2025-10-11 03:56:42.234633236 +0000 UTC m=+3049.793578021" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.576477 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.663861 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w8fv\" (UniqueName: \"kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv\") pod \"f7055e1d-ffb5-4cca-9a53-2400675059a0\" (UID: \"f7055e1d-ffb5-4cca-9a53-2400675059a0\") " Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.672339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv" (OuterVolumeSpecName: "kube-api-access-9w8fv") pod "f7055e1d-ffb5-4cca-9a53-2400675059a0" (UID: "f7055e1d-ffb5-4cca-9a53-2400675059a0"). InnerVolumeSpecName "kube-api-access-9w8fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:56:42 crc kubenswrapper[4754]: I1011 03:56:42.766777 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w8fv\" (UniqueName: \"kubernetes.io/projected/f7055e1d-ffb5-4cca-9a53-2400675059a0-kube-api-access-9w8fv\") on node \"crc\" DevicePath \"\"" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.125918 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ed00-account-create-q2rgc" event={"ID":"f7055e1d-ffb5-4cca-9a53-2400675059a0","Type":"ContainerDied","Data":"06be9f75840c36c64b84cebba80c3698c31f45bbaa1def5e09a36bd6a90b3c7c"} Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.126023 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06be9f75840c36c64b84cebba80c3698c31f45bbaa1def5e09a36bd6a90b3c7c" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.126365 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ed00-account-create-q2rgc" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.407485 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.408607 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.421809 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.421857 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.452759 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.452832 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.454518 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:43 crc kubenswrapper[4754]: I1011 03:56:43.486515 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:44 crc kubenswrapper[4754]: I1011 03:56:44.137390 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:44 crc kubenswrapper[4754]: I1011 03:56:44.137715 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 03:56:44 crc kubenswrapper[4754]: I1011 03:56:44.137785 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:44 crc kubenswrapper[4754]: I1011 03:56:44.137809 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 03:56:46 crc kubenswrapper[4754]: I1011 03:56:46.376515 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 03:56:46 crc kubenswrapper[4754]: I1011 03:56:46.456703 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:46 crc kubenswrapper[4754]: I1011 03:56:46.473588 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.127204 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-kpkwx"] Oct 11 03:56:47 crc kubenswrapper[4754]: E1011 03:56:47.127590 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7055e1d-ffb5-4cca-9a53-2400675059a0" containerName="mariadb-account-create" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.127602 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7055e1d-ffb5-4cca-9a53-2400675059a0" containerName="mariadb-account-create" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.127823 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7055e1d-ffb5-4cca-9a53-2400675059a0" containerName="mariadb-account-create" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.128449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.130410 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7xp82" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.139564 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.166213 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-kpkwx"] Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.181978 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.318936 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p8zh\" (UniqueName: \"kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.319030 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.319084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.319118 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.398552 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.422036 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.422106 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.422256 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p8zh\" (UniqueName: \"kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.422286 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.431463 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.441116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.443725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.446755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p8zh\" (UniqueName: \"kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh\") pod \"manila-db-sync-kpkwx\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:47 crc kubenswrapper[4754]: I1011 03:56:47.448664 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-kpkwx" Oct 11 03:56:48 crc kubenswrapper[4754]: I1011 03:56:48.108429 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-kpkwx"] Oct 11 03:56:48 crc kubenswrapper[4754]: I1011 03:56:48.190227 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-kpkwx" event={"ID":"5e2dd8cc-cbb0-4675-b9a4-46846b783730","Type":"ContainerStarted","Data":"4570e2d6d3d1f0829f31efd3f960a0f7875001f98099142c158aee24c7ef5ef7"} Oct 11 03:56:48 crc kubenswrapper[4754]: I1011 03:56:48.642307 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 03:56:50 crc kubenswrapper[4754]: I1011 03:56:50.144205 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:50 crc kubenswrapper[4754]: I1011 03:56:50.144564 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:56:50 crc kubenswrapper[4754]: I1011 03:56:50.318423 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:50 crc kubenswrapper[4754]: I1011 03:56:50.318499 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:56:53 crc kubenswrapper[4754]: I1011 03:56:53.239433 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-kpkwx" event={"ID":"5e2dd8cc-cbb0-4675-b9a4-46846b783730","Type":"ContainerStarted","Data":"f615f38466605f507c3cc08ba551cc10f5513934ae868e731628053eee2d1c3c"} Oct 11 03:56:53 crc kubenswrapper[4754]: I1011 03:56:53.267654 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-kpkwx" podStartSLOduration=1.872193022 podStartE2EDuration="6.26762523s" podCreationTimestamp="2025-10-11 03:56:47 +0000 UTC" firstStartedPulling="2025-10-11 03:56:48.13564027 +0000 UTC m=+3055.694585045" lastFinishedPulling="2025-10-11 03:56:52.531072468 +0000 UTC m=+3060.090017253" observedRunningTime="2025-10-11 03:56:53.257006072 +0000 UTC m=+3060.815950897" watchObservedRunningTime="2025-10-11 03:56:53.26762523 +0000 UTC m=+3060.826570035" Oct 11 03:57:00 crc kubenswrapper[4754]: I1011 03:57:00.146635 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.239:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.239:8443: connect: connection refused" Oct 11 03:57:00 crc kubenswrapper[4754]: I1011 03:57:00.319206 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8655c657c6-4mw6n" podUID="96d0a63b-15e8-4830-8fd6-1bd47e6ed973" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Oct 11 03:57:00 crc kubenswrapper[4754]: I1011 03:57:00.737135 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:57:00 crc kubenswrapper[4754]: I1011 03:57:00.737238 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:57:02 crc kubenswrapper[4754]: I1011 03:57:02.345173 4754 generic.go:334] "Generic (PLEG): container finished" podID="5e2dd8cc-cbb0-4675-b9a4-46846b783730" containerID="f615f38466605f507c3cc08ba551cc10f5513934ae868e731628053eee2d1c3c" exitCode=0 Oct 11 03:57:02 crc kubenswrapper[4754]: I1011 03:57:02.345252 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-kpkwx" event={"ID":"5e2dd8cc-cbb0-4675-b9a4-46846b783730","Type":"ContainerDied","Data":"f615f38466605f507c3cc08ba551cc10f5513934ae868e731628053eee2d1c3c"} Oct 11 03:57:03 crc kubenswrapper[4754]: I1011 03:57:03.892684 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-kpkwx" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.026078 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data\") pod \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.028014 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p8zh\" (UniqueName: \"kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh\") pod \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.028307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data\") pod \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.028478 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle\") pod \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\" (UID: \"5e2dd8cc-cbb0-4675-b9a4-46846b783730\") " Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.033674 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh" (OuterVolumeSpecName: "kube-api-access-5p8zh") pod "5e2dd8cc-cbb0-4675-b9a4-46846b783730" (UID: "5e2dd8cc-cbb0-4675-b9a4-46846b783730"). InnerVolumeSpecName "kube-api-access-5p8zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.034305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "5e2dd8cc-cbb0-4675-b9a4-46846b783730" (UID: "5e2dd8cc-cbb0-4675-b9a4-46846b783730"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.041031 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data" (OuterVolumeSpecName: "config-data") pod "5e2dd8cc-cbb0-4675-b9a4-46846b783730" (UID: "5e2dd8cc-cbb0-4675-b9a4-46846b783730"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.067507 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e2dd8cc-cbb0-4675-b9a4-46846b783730" (UID: "5e2dd8cc-cbb0-4675-b9a4-46846b783730"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.132170 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p8zh\" (UniqueName: \"kubernetes.io/projected/5e2dd8cc-cbb0-4675-b9a4-46846b783730-kube-api-access-5p8zh\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.132633 4754 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.132764 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.132911 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2dd8cc-cbb0-4675-b9a4-46846b783730-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.389220 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-kpkwx" event={"ID":"5e2dd8cc-cbb0-4675-b9a4-46846b783730","Type":"ContainerDied","Data":"4570e2d6d3d1f0829f31efd3f960a0f7875001f98099142c158aee24c7ef5ef7"} Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.389282 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4570e2d6d3d1f0829f31efd3f960a0f7875001f98099142c158aee24c7ef5ef7" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.389373 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-kpkwx" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.838679 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:04 crc kubenswrapper[4754]: E1011 03:57:04.839701 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2dd8cc-cbb0-4675-b9a4-46846b783730" containerName="manila-db-sync" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.839723 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2dd8cc-cbb0-4675-b9a4-46846b783730" containerName="manila-db-sync" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.840051 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2dd8cc-cbb0-4675-b9a4-46846b783730" containerName="manila-db-sync" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.841383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.849682 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.849984 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7xp82" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.850116 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.850291 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.854826 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.890249 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.892367 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.900458 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.917036 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956188 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956246 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956291 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqg2t\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956430 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twzh8\" (UniqueName: \"kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956511 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.956982 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:04 crc kubenswrapper[4754]: I1011 03:57:04.957068 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.002997 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pz4xn"] Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.005354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.021244 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pz4xn"] Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065659 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065741 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065804 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065874 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqg2t\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.065959 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twzh8\" (UniqueName: \"kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066304 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066365 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066408 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066473 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066599 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.066675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.091071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.091374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.093293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.098936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.107979 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.123192 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.125009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.125380 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.137194 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqg2t\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.138836 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.140492 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.135280 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.146017 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.164373 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twzh8\" (UniqueName: \"kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8\") pod \"manila-scheduler-0\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188202 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188268 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-config\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188411 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r28nk\" (UniqueName: \"kubernetes.io/projected/f3601d47-a321-4202-9813-6b9ce785cbee-kube-api-access-r28nk\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.188471 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.196331 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.198408 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.198502 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.201023 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.215559 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.237867 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.294068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-config\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.294700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r28nk\" (UniqueName: \"kubernetes.io/projected/f3601d47-a321-4202-9813-6b9ce785cbee-kube-api-access-r28nk\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295239 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-config\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295466 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295628 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.295831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.308275 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.311227 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.315091 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f3601d47-a321-4202-9813-6b9ce785cbee-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.324654 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r28nk\" (UniqueName: \"kubernetes.io/projected/f3601d47-a321-4202-9813-6b9ce785cbee-kube-api-access-r28nk\") pod \"dnsmasq-dns-76b5fdb995-pz4xn\" (UID: \"f3601d47-a321-4202-9813-6b9ce785cbee\") " pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.341439 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.400793 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.400854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7cfk\" (UniqueName: \"kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.400897 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.400923 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.400950 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.401173 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:05 crc kubenswrapper[4754]: I1011 03:57:05.401252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.506626 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507109 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507369 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7cfk\" (UniqueName: \"kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507441 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.507470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.508379 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.508935 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.530011 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.530562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.544729 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.546859 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.558748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7cfk\" (UniqueName: \"kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk\") pod \"manila-api-0\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:05.848955 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:06.554979 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:06.661692 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pz4xn"] Oct 11 03:57:06 crc kubenswrapper[4754]: I1011 03:57:06.772575 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.471574 4754 generic.go:334] "Generic (PLEG): container finished" podID="f3601d47-a321-4202-9813-6b9ce785cbee" containerID="bc9de749f74a42b4f96b885233b6c971e66477a9fc71fc674a67d3f66e468cba" exitCode=0 Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.471804 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" event={"ID":"f3601d47-a321-4202-9813-6b9ce785cbee","Type":"ContainerDied","Data":"bc9de749f74a42b4f96b885233b6c971e66477a9fc71fc674a67d3f66e468cba"} Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.472434 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" event={"ID":"f3601d47-a321-4202-9813-6b9ce785cbee","Type":"ContainerStarted","Data":"103a8c784fb0d18849fc88a9f8029bcccb26b7241d57005559bd1634474748b9"} Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.482436 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerStarted","Data":"1c894501d8756a6fa67fad90677cac92da1ef3ce2d5d338fcce99fd13663d8f4"} Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.485010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerStarted","Data":"f1a2b3f87e8c78a7cade995140ca884a14ee8106ff35252200a64df8c0feeb36"} Oct 11 03:57:07 crc kubenswrapper[4754]: I1011 03:57:07.893241 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.241303 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.530363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" event={"ID":"f3601d47-a321-4202-9813-6b9ce785cbee","Type":"ContainerStarted","Data":"02db6d47e49deed7ceb8f949363540ba3225e9682f8a7b824c35747c3959bb1d"} Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.535649 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.546547 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerStarted","Data":"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383"} Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.546639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerStarted","Data":"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211"} Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.550573 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerStarted","Data":"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300"} Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.550626 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerStarted","Data":"959c5fc5e506af27b1c6c113a1dbb3beead6a02b61965d4e6817992a873d70af"} Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.598585 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" podStartSLOduration=4.598549191 podStartE2EDuration="4.598549191s" podCreationTimestamp="2025-10-11 03:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:57:08.572330723 +0000 UTC m=+3076.131275508" watchObservedRunningTime="2025-10-11 03:57:08.598549191 +0000 UTC m=+3076.157493976" Oct 11 03:57:08 crc kubenswrapper[4754]: I1011 03:57:08.605782 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.894450651 podStartE2EDuration="4.605755283s" podCreationTimestamp="2025-10-11 03:57:04 +0000 UTC" firstStartedPulling="2025-10-11 03:57:06.569948957 +0000 UTC m=+3074.128893742" lastFinishedPulling="2025-10-11 03:57:07.281253589 +0000 UTC m=+3074.840198374" observedRunningTime="2025-10-11 03:57:08.591654857 +0000 UTC m=+3076.150599652" watchObservedRunningTime="2025-10-11 03:57:08.605755283 +0000 UTC m=+3076.164700058" Oct 11 03:57:09 crc kubenswrapper[4754]: I1011 03:57:09.567859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerStarted","Data":"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69"} Oct 11 03:57:09 crc kubenswrapper[4754]: I1011 03:57:09.568804 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api-log" containerID="cri-o://123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" gracePeriod=30 Oct 11 03:57:09 crc kubenswrapper[4754]: I1011 03:57:09.569449 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api" containerID="cri-o://13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" gracePeriod=30 Oct 11 03:57:09 crc kubenswrapper[4754]: I1011 03:57:09.603709 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.6036793750000005 podStartE2EDuration="4.603679375s" podCreationTimestamp="2025-10-11 03:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:57:09.593162749 +0000 UTC m=+3077.152107534" watchObservedRunningTime="2025-10-11 03:57:09.603679375 +0000 UTC m=+3077.162624150" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.290332 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.369937 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.370289 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.376529 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs" (OuterVolumeSpecName: "logs") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.376621 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.376653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.376825 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.376910 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.377076 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7cfk\" (UniqueName: \"kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk\") pod \"75ebe333-9aae-4efd-a724-65833a3eba02\" (UID: \"75ebe333-9aae-4efd-a724-65833a3eba02\") " Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.379443 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ebe333-9aae-4efd-a724-65833a3eba02-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.385398 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.385665 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts" (OuterVolumeSpecName: "scripts") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.387658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk" (OuterVolumeSpecName: "kube-api-access-q7cfk") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "kube-api-access-q7cfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.410202 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.435015 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.467412 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data" (OuterVolumeSpecName: "config-data") pod "75ebe333-9aae-4efd-a724-65833a3eba02" (UID: "75ebe333-9aae-4efd-a724-65833a3eba02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481673 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7cfk\" (UniqueName: \"kubernetes.io/projected/75ebe333-9aae-4efd-a724-65833a3eba02-kube-api-access-q7cfk\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481711 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481722 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75ebe333-9aae-4efd-a724-65833a3eba02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481734 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481743 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.481753 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75ebe333-9aae-4efd-a724-65833a3eba02-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.569572 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.569945 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-central-agent" containerID="cri-o://f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592" gracePeriod=30 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.570547 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="proxy-httpd" containerID="cri-o://27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06" gracePeriod=30 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.570639 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-notification-agent" containerID="cri-o://38127e3f1de4bafc2552a6c1b37bf8bdfb882b7b76734b132471363d941e2efb" gracePeriod=30 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.570626 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="sg-core" containerID="cri-o://669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903" gracePeriod=30 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595832 4754 generic.go:334] "Generic (PLEG): container finished" podID="75ebe333-9aae-4efd-a724-65833a3eba02" containerID="13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" exitCode=0 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595869 4754 generic.go:334] "Generic (PLEG): container finished" podID="75ebe333-9aae-4efd-a724-65833a3eba02" containerID="123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" exitCode=143 Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerDied","Data":"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69"} Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerDied","Data":"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300"} Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75ebe333-9aae-4efd-a724-65833a3eba02","Type":"ContainerDied","Data":"959c5fc5e506af27b1c6c113a1dbb3beead6a02b61965d4e6817992a873d70af"} Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.595956 4754 scope.go:117] "RemoveContainer" containerID="13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.596157 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.644317 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.658901 4754 scope.go:117] "RemoveContainer" containerID="123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.667875 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.688743 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:10 crc kubenswrapper[4754]: E1011 03:57:10.689343 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.689362 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api" Oct 11 03:57:10 crc kubenswrapper[4754]: E1011 03:57:10.689387 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api-log" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.689394 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api-log" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.689630 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.689650 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" containerName="manila-api-log" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.691403 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.695507 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.695759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.695661 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.713422 4754 scope.go:117] "RemoveContainer" containerID="13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" Oct 11 03:57:10 crc kubenswrapper[4754]: E1011 03:57:10.715785 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69\": container with ID starting with 13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69 not found: ID does not exist" containerID="13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.715821 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69"} err="failed to get container status \"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69\": rpc error: code = NotFound desc = could not find container \"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69\": container with ID starting with 13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69 not found: ID does not exist" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.715842 4754 scope.go:117] "RemoveContainer" containerID="123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" Oct 11 03:57:10 crc kubenswrapper[4754]: E1011 03:57:10.720848 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300\": container with ID starting with 123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300 not found: ID does not exist" containerID="123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.720932 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300"} err="failed to get container status \"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300\": rpc error: code = NotFound desc = could not find container \"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300\": container with ID starting with 123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300 not found: ID does not exist" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.721017 4754 scope.go:117] "RemoveContainer" containerID="13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.721476 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.727171 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69"} err="failed to get container status \"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69\": rpc error: code = NotFound desc = could not find container \"13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69\": container with ID starting with 13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69 not found: ID does not exist" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.727222 4754 scope.go:117] "RemoveContainer" containerID="123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.733168 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300"} err="failed to get container status \"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300\": rpc error: code = NotFound desc = could not find container \"123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300\": container with ID starting with 123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300 not found: ID does not exist" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aeef235-4c59-4b53-bb5a-5284a5c3066b-etc-machine-id\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789191 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789232 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-scripts\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789261 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789302 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwt8d\" (UniqueName: \"kubernetes.io/projected/0aeef235-4c59-4b53-bb5a-5284a5c3066b-kube-api-access-mwt8d\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789328 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aeef235-4c59-4b53-bb5a-5284a5c3066b-logs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789371 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-public-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789457 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data-custom\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.789529 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwt8d\" (UniqueName: \"kubernetes.io/projected/0aeef235-4c59-4b53-bb5a-5284a5c3066b-kube-api-access-mwt8d\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aeef235-4c59-4b53-bb5a-5284a5c3066b-logs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-public-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891234 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data-custom\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891290 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891322 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aeef235-4c59-4b53-bb5a-5284a5c3066b-etc-machine-id\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891339 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891365 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-scripts\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.891382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.900182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aeef235-4c59-4b53-bb5a-5284a5c3066b-etc-machine-id\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.900731 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aeef235-4c59-4b53-bb5a-5284a5c3066b-logs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.901294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.901472 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.901760 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-scripts\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.904521 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.908714 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-public-tls-certs\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.912127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aeef235-4c59-4b53-bb5a-5284a5c3066b-config-data-custom\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:10 crc kubenswrapper[4754]: I1011 03:57:10.922730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwt8d\" (UniqueName: \"kubernetes.io/projected/0aeef235-4c59-4b53-bb5a-5284a5c3066b-kube-api-access-mwt8d\") pod \"manila-api-0\" (UID: \"0aeef235-4c59-4b53-bb5a-5284a5c3066b\") " pod="openstack/manila-api-0" Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.111272 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75ebe333-9aae-4efd-a724-65833a3eba02" path="/var/lib/kubelet/pods/75ebe333-9aae-4efd-a724-65833a3eba02/volumes" Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.151469 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.617834 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerID="27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06" exitCode=0 Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.618278 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerDied","Data":"27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06"} Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.618292 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerID="669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903" exitCode=2 Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.618307 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerID="f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592" exitCode=0 Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.618311 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerDied","Data":"669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903"} Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.618322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerDied","Data":"f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592"} Oct 11 03:57:11 crc kubenswrapper[4754]: I1011 03:57:11.805358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 11 03:57:12 crc kubenswrapper[4754]: W1011 03:57:12.194627 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-959c5fc5e506af27b1c6c113a1dbb3beead6a02b61965d4e6817992a873d70af": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-959c5fc5e506af27b1c6c113a1dbb3beead6a02b61965d4e6817992a873d70af: no such file or directory Oct 11 03:57:12 crc kubenswrapper[4754]: W1011 03:57:12.195259 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-conmon-123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-conmon-123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300.scope: no such file or directory Oct 11 03:57:12 crc kubenswrapper[4754]: W1011 03:57:12.195388 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-123f705ed7567993151e92eccb354cf4aa09dd38613b9ef1f96dd8ee90bba300.scope: no such file or directory Oct 11 03:57:12 crc kubenswrapper[4754]: W1011 03:57:12.195468 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-conmon-13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-conmon-13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69.scope: no such file or directory Oct 11 03:57:12 crc kubenswrapper[4754]: W1011 03:57:12.195488 4754 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice/crio-13396f6917a3ffc58a817e7fafb8be67a591eccbd438c4ffa3fb7d36381cac69.scope: no such file or directory Oct 11 03:57:12 crc kubenswrapper[4754]: E1011 03:57:12.551683 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd085f1b_5d07_443b_a213_3a64678bb76a.slice/crio-conmon-3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a35def2_232b_4352_9e1b_0d13b32bbc3e.slice/crio-27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a35def2_232b_4352_9e1b_0d13b32bbc3e.slice/crio-f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a35def2_232b_4352_9e1b_0d13b32bbc3e.slice/crio-conmon-669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c190463_553c_41d4_ac6c_122832a2b286.slice/crio-conmon-8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a35def2_232b_4352_9e1b_0d13b32bbc3e.slice/crio-conmon-f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c190463_553c_41d4_ac6c_122832a2b286.slice/crio-88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ebe333_9aae_4efd_a724_65833a3eba02.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c190463_553c_41d4_ac6c_122832a2b286.slice/crio-8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd085f1b_5d07_443b_a213_3a64678bb76a.slice/crio-ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3601d47_a321_4202_9813_6b9ce785cbee.slice/crio-conmon-bc9de749f74a42b4f96b885233b6c971e66477a9fc71fc674a67d3f66e468cba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd085f1b_5d07_443b_a213_3a64678bb76a.slice/crio-conmon-ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3601d47_a321_4202_9813_6b9ce785cbee.slice/crio-bc9de749f74a42b4f96b885233b6c971e66477a9fc71fc674a67d3f66e468cba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a35def2_232b_4352_9e1b_0d13b32bbc3e.slice/crio-669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd085f1b_5d07_443b_a213_3a64678bb76a.slice/crio-3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c190463_553c_41d4_ac6c_122832a2b286.slice/crio-conmon-88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.670274 4754 generic.go:334] "Generic (PLEG): container finished" podID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerID="3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13" exitCode=137 Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.670343 4754 generic.go:334] "Generic (PLEG): container finished" podID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerID="ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667" exitCode=137 Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.670335 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerDied","Data":"3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.670468 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerDied","Data":"ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.672618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0aeef235-4c59-4b53-bb5a-5284a5c3066b","Type":"ContainerStarted","Data":"27f4d2e57a2d755db93bdb70d624045712cb6087dc0eb838eaf036ad5c96d30e"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.672645 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0aeef235-4c59-4b53-bb5a-5284a5c3066b","Type":"ContainerStarted","Data":"b5d2572de41eec28bd2b3b7fd139b0ea6804db5cda4126818a9ce64c14844914"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.674946 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c190463-553c-41d4-ac6c-122832a2b286" containerID="8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07" exitCode=137 Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.675044 4754 generic.go:334] "Generic (PLEG): container finished" podID="0c190463-553c-41d4-ac6c-122832a2b286" containerID="88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe" exitCode=137 Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.675156 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerDied","Data":"8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.675189 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerDied","Data":"88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe"} Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.702502 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:57:12 crc kubenswrapper[4754]: I1011 03:57:12.738642 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:57:13 crc kubenswrapper[4754]: I1011 03:57:13.698565 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerID="38127e3f1de4bafc2552a6c1b37bf8bdfb882b7b76734b132471363d941e2efb" exitCode=0 Oct 11 03:57:13 crc kubenswrapper[4754]: I1011 03:57:13.698943 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerDied","Data":"38127e3f1de4bafc2552a6c1b37bf8bdfb882b7b76734b132471363d941e2efb"} Oct 11 03:57:14 crc kubenswrapper[4754]: I1011 03:57:14.762743 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8655c657c6-4mw6n" Oct 11 03:57:14 crc kubenswrapper[4754]: I1011 03:57:14.772346 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:57:14 crc kubenswrapper[4754]: I1011 03:57:14.841583 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.202588 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.344011 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b5fdb995-pz4xn" Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.415056 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.415378 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="dnsmasq-dns" containerID="cri-o://6f1d86e44f592c3e7d660121f5138a79c37b9c4478384c5e4e2532841acf6310" gracePeriod=10 Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.727276 4754 generic.go:334] "Generic (PLEG): container finished" podID="e6098849-4d4b-4519-8e0e-548914f42c85" containerID="6f1d86e44f592c3e7d660121f5138a79c37b9c4478384c5e4e2532841acf6310" exitCode=0 Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.727570 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon-log" containerID="cri-o://fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121" gracePeriod=30 Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.727694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" event={"ID":"e6098849-4d4b-4519-8e0e-548914f42c85","Type":"ContainerDied","Data":"6f1d86e44f592c3e7d660121f5138a79c37b9c4478384c5e4e2532841acf6310"} Oct 11 03:57:15 crc kubenswrapper[4754]: I1011 03:57:15.728303 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" containerID="cri-o://173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e" gracePeriod=30 Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.669350 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.674255 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.686462 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data\") pod \"fd085f1b-5d07-443b-a213-3a64678bb76a\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.686693 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key\") pod \"fd085f1b-5d07-443b-a213-3a64678bb76a\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.686944 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs\") pod \"0c190463-553c-41d4-ac6c-122832a2b286\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687021 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8lpt\" (UniqueName: \"kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt\") pod \"0c190463-553c-41d4-ac6c-122832a2b286\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687074 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8jtl\" (UniqueName: \"kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl\") pod \"fd085f1b-5d07-443b-a213-3a64678bb76a\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687144 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts\") pod \"fd085f1b-5d07-443b-a213-3a64678bb76a\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key\") pod \"0c190463-553c-41d4-ac6c-122832a2b286\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687230 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts\") pod \"0c190463-553c-41d4-ac6c-122832a2b286\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data\") pod \"0c190463-553c-41d4-ac6c-122832a2b286\" (UID: \"0c190463-553c-41d4-ac6c-122832a2b286\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.687370 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs\") pod \"fd085f1b-5d07-443b-a213-3a64678bb76a\" (UID: \"fd085f1b-5d07-443b-a213-3a64678bb76a\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.689277 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs" (OuterVolumeSpecName: "logs") pod "0c190463-553c-41d4-ac6c-122832a2b286" (UID: "0c190463-553c-41d4-ac6c-122832a2b286"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.690035 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs" (OuterVolumeSpecName: "logs") pod "fd085f1b-5d07-443b-a213-3a64678bb76a" (UID: "fd085f1b-5d07-443b-a213-3a64678bb76a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.701084 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c190463-553c-41d4-ac6c-122832a2b286-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.701115 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd085f1b-5d07-443b-a213-3a64678bb76a-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.730145 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt" (OuterVolumeSpecName: "kube-api-access-g8lpt") pod "0c190463-553c-41d4-ac6c-122832a2b286" (UID: "0c190463-553c-41d4-ac6c-122832a2b286"). InnerVolumeSpecName "kube-api-access-g8lpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.737763 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl" (OuterVolumeSpecName: "kube-api-access-p8jtl") pod "fd085f1b-5d07-443b-a213-3a64678bb76a" (UID: "fd085f1b-5d07-443b-a213-3a64678bb76a"). InnerVolumeSpecName "kube-api-access-p8jtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.738282 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fd085f1b-5d07-443b-a213-3a64678bb76a" (UID: "fd085f1b-5d07-443b-a213-3a64678bb76a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.738675 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts" (OuterVolumeSpecName: "scripts") pod "fd085f1b-5d07-443b-a213-3a64678bb76a" (UID: "fd085f1b-5d07-443b-a213-3a64678bb76a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.739299 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0c190463-553c-41d4-ac6c-122832a2b286" (UID: "0c190463-553c-41d4-ac6c-122832a2b286"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.746639 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts" (OuterVolumeSpecName: "scripts") pod "0c190463-553c-41d4-ac6c-122832a2b286" (UID: "0c190463-553c-41d4-ac6c-122832a2b286"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.759066 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data" (OuterVolumeSpecName: "config-data") pod "fd085f1b-5d07-443b-a213-3a64678bb76a" (UID: "fd085f1b-5d07-443b-a213-3a64678bb76a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.768668 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data" (OuterVolumeSpecName: "config-data") pod "0c190463-553c-41d4-ac6c-122832a2b286" (UID: "0c190463-553c-41d4-ac6c-122832a2b286"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.770351 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-785c858b67-vvjnn" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.770572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-785c858b67-vvjnn" event={"ID":"0c190463-553c-41d4-ac6c-122832a2b286","Type":"ContainerDied","Data":"fa4384dff4d5eac69ad3c4833fb8d0fd3fc3e3a03f0b1745c369c1674229ac90"} Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.770649 4754 scope.go:117] "RemoveContainer" containerID="8950db5598041290204bbf76326b8abe946c9bdcbe8c509d2b0ea074dd302a07" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.776364 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86df4d8c9c-qtbx4" event={"ID":"fd085f1b-5d07-443b-a213-3a64678bb76a","Type":"ContainerDied","Data":"d535bdeb161598f52260227f909eff5e52776c7a7268d00df21db5a2b3fadf5c"} Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.776479 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86df4d8c9c-qtbx4" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803453 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803483 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fd085f1b-5d07-443b-a213-3a64678bb76a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803497 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8lpt\" (UniqueName: \"kubernetes.io/projected/0c190463-553c-41d4-ac6c-122832a2b286-kube-api-access-g8lpt\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803509 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8jtl\" (UniqueName: \"kubernetes.io/projected/fd085f1b-5d07-443b-a213-3a64678bb76a-kube-api-access-p8jtl\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803524 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd085f1b-5d07-443b-a213-3a64678bb76a-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803534 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c190463-553c-41d4-ac6c-122832a2b286-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803543 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.803552 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c190463-553c-41d4-ac6c-122832a2b286-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.817956 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.903031 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.908067 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.908334 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.908423 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.909109 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.909265 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.909478 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.909568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmkbk\" (UniqueName: \"kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.909633 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data\") pod \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\" (UID: \"5a35def2-232b-4352-9e1b-0d13b32bbc3e\") " Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.922651 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.923043 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.923422 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-785c858b67-vvjnn"] Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.930667 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk" (OuterVolumeSpecName: "kube-api-access-rmkbk") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "kube-api-access-rmkbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.933727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts" (OuterVolumeSpecName: "scripts") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.945601 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.959708 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86df4d8c9c-qtbx4"] Oct 11 03:57:16 crc kubenswrapper[4754]: I1011 03:57:16.996071 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.012553 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.012601 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmkbk\" (UniqueName: \"kubernetes.io/projected/5a35def2-232b-4352-9e1b-0d13b32bbc3e-kube-api-access-rmkbk\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.012617 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.012630 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.012642 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a35def2-232b-4352-9e1b-0d13b32bbc3e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.022392 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.099076 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c190463-553c-41d4-ac6c-122832a2b286" path="/var/lib/kubelet/pods/0c190463-553c-41d4-ac6c-122832a2b286/volumes" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.100424 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" path="/var/lib/kubelet/pods/fd085f1b-5d07-443b-a213-3a64678bb76a/volumes" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.114784 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.115095 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.118270 4754 scope.go:117] "RemoveContainer" containerID="88551ce0e1adbb1370fb020b1a71017ee8eede959eb5e57c739f4c9279ec14fe" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.145594 4754 scope.go:117] "RemoveContainer" containerID="3ab7b368be8b47b5db2967e1680c331e2c5e52a1ac1146f21f4729fdaa4b6d13" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.149704 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data" (OuterVolumeSpecName: "config-data") pod "5a35def2-232b-4352-9e1b-0d13b32bbc3e" (UID: "5a35def2-232b-4352-9e1b-0d13b32bbc3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.161193 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.215729 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.215834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.215891 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.215993 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.216167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb2tw\" (UniqueName: \"kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.216203 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config\") pod \"e6098849-4d4b-4519-8e0e-548914f42c85\" (UID: \"e6098849-4d4b-4519-8e0e-548914f42c85\") " Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.216831 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.216857 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a35def2-232b-4352-9e1b-0d13b32bbc3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.232091 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw" (OuterVolumeSpecName: "kube-api-access-hb2tw") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "kube-api-access-hb2tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.277991 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.283367 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config" (OuterVolumeSpecName: "config") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.288167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.300478 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.311749 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6098849-4d4b-4519-8e0e-548914f42c85" (UID: "e6098849-4d4b-4519-8e0e-548914f42c85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.321887 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.321936 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.321946 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.321957 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb2tw\" (UniqueName: \"kubernetes.io/projected/e6098849-4d4b-4519-8e0e-548914f42c85-kube-api-access-hb2tw\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.322058 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.322069 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e6098849-4d4b-4519-8e0e-548914f42c85-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.338371 4754 scope.go:117] "RemoveContainer" containerID="ad5a7de9a7d729a7ad435e526c575fb3a2edfb9e15fed70b986ef3c2db4a0667" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.792597 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0aeef235-4c59-4b53-bb5a-5284a5c3066b","Type":"ContainerStarted","Data":"ca0fa65a981352b0dd07f18c61834aea4821fd465d5a07f51da9f01e330cd20d"} Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.794984 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.803153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" event={"ID":"e6098849-4d4b-4519-8e0e-548914f42c85","Type":"ContainerDied","Data":"0701c77c05bdecf3682c2b3cc2647c4c02fca10901d173b98f2d1d54cda31224"} Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.803202 4754 scope.go:117] "RemoveContainer" containerID="6f1d86e44f592c3e7d660121f5138a79c37b9c4478384c5e4e2532841acf6310" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.803323 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-879wl" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.828580 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=7.828551346 podStartE2EDuration="7.828551346s" podCreationTimestamp="2025-10-11 03:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:57:17.824855142 +0000 UTC m=+3085.383799927" watchObservedRunningTime="2025-10-11 03:57:17.828551346 +0000 UTC m=+3085.387496131" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.842692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a35def2-232b-4352-9e1b-0d13b32bbc3e","Type":"ContainerDied","Data":"6a8f61075043d247e92311c189fa15b632341c512b4a5b65e9cd09e4bb1b7d7c"} Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.842842 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.861195 4754 scope.go:117] "RemoveContainer" containerID="41995a927cc6e9d1f0eaefab46648296e3bd41164adf1f9970e6b4c99b13c406" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.862485 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerStarted","Data":"644ca25a9d632e868c3f1e02b7cc04817e1d4f77f3814715b396360a240f4de7"} Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.864007 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.881529 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-879wl"] Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.904054 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.907550 4754 scope.go:117] "RemoveContainer" containerID="27ac9705b8894e0440b25d1049f8d1b76b1df8bb69abebdbf405bd7183b3cb06" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.914711 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.928806 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930616 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="init" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930646 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="init" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930666 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930676 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930693 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="proxy-httpd" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930706 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="proxy-httpd" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930719 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930727 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930739 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="dnsmasq-dns" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930747 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="dnsmasq-dns" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930778 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-central-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930787 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-central-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930804 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930812 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930833 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="sg-core" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930841 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="sg-core" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930858 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930866 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: E1011 03:57:17.930880 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-notification-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.930890 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-notification-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931584 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="proxy-httpd" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931613 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931637 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" containerName="dnsmasq-dns" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931656 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931669 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd085f1b-5d07-443b-a213-3a64678bb76a" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931679 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c190463-553c-41d4-ac6c-122832a2b286" containerName="horizon-log" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931693 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-notification-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931736 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="ceilometer-central-agent" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.931747 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" containerName="sg-core" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.937188 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.940560 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.940622 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.941027 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.944617 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:17 crc kubenswrapper[4754]: I1011 03:57:17.985434 4754 scope.go:117] "RemoveContainer" containerID="669c60401c2939d639e314e2e46c080219afdab8335907b8a112a001b5740903" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.007214 4754 scope.go:117] "RemoveContainer" containerID="38127e3f1de4bafc2552a6c1b37bf8bdfb882b7b76734b132471363d941e2efb" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.036092 4754 scope.go:117] "RemoveContainer" containerID="f3c0bb80edae67d9f2736f1bae2cd447b452b21d686836e4e0f39d7621986592" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056042 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056123 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056435 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056600 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw2ss\" (UniqueName: \"kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.056763 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.158758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.158889 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.159002 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.160467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.160536 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw2ss\" (UniqueName: \"kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.160811 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.160889 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.161123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.161101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.161846 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.168992 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.171095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.172473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.174879 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.179653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.181589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw2ss\" (UniqueName: \"kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss\") pod \"ceilometer-0\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.284385 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.790668 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.875422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerStarted","Data":"a1c541bf1f4b46c2b2af1d36302b6bbbf0447290f5a792c1003b2ceabb97a114"} Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.885110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerStarted","Data":"c68f8833329e3293a0fe38e7b4be65ecd166fb3f8c585e7163e5481a07560123"} Oct 11 03:57:18 crc kubenswrapper[4754]: I1011 03:57:18.907584 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=5.211871018 podStartE2EDuration="14.907563338s" podCreationTimestamp="2025-10-11 03:57:04 +0000 UTC" firstStartedPulling="2025-10-11 03:57:06.778254375 +0000 UTC m=+3074.337199160" lastFinishedPulling="2025-10-11 03:57:16.473946705 +0000 UTC m=+3084.032891480" observedRunningTime="2025-10-11 03:57:18.904528142 +0000 UTC m=+3086.463472947" watchObservedRunningTime="2025-10-11 03:57:18.907563338 +0000 UTC m=+3086.466508123" Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.113681 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a35def2-232b-4352-9e1b-0d13b32bbc3e" path="/var/lib/kubelet/pods/5a35def2-232b-4352-9e1b-0d13b32bbc3e/volumes" Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.120935 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6098849-4d4b-4519-8e0e-548914f42c85" path="/var/lib/kubelet/pods/e6098849-4d4b-4519-8e0e-548914f42c85/volumes" Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.344129 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.896264 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerID="173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e" exitCode=0 Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.896350 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerDied","Data":"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e"} Oct 11 03:57:19 crc kubenswrapper[4754]: I1011 03:57:19.898201 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerStarted","Data":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} Oct 11 03:57:20 crc kubenswrapper[4754]: I1011 03:57:20.145651 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.239:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.239:8443: connect: connection refused" Oct 11 03:57:20 crc kubenswrapper[4754]: I1011 03:57:20.911093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerStarted","Data":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} Oct 11 03:57:21 crc kubenswrapper[4754]: I1011 03:57:21.923246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerStarted","Data":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.946850 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerStarted","Data":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.947200 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-central-agent" containerID="cri-o://8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" gracePeriod=30 Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.947435 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.947466 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="proxy-httpd" containerID="cri-o://db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" gracePeriod=30 Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.947523 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="sg-core" containerID="cri-o://f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" gracePeriod=30 Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.947559 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-notification-agent" containerID="cri-o://20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" gracePeriod=30 Oct 11 03:57:23 crc kubenswrapper[4754]: I1011 03:57:23.983724 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.813145233 podStartE2EDuration="6.983704247s" podCreationTimestamp="2025-10-11 03:57:17 +0000 UTC" firstStartedPulling="2025-10-11 03:57:18.805087126 +0000 UTC m=+3086.364031911" lastFinishedPulling="2025-10-11 03:57:22.97564614 +0000 UTC m=+3090.534590925" observedRunningTime="2025-10-11 03:57:23.976747891 +0000 UTC m=+3091.535692666" watchObservedRunningTime="2025-10-11 03:57:23.983704247 +0000 UTC m=+3091.542649032" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.855894 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.951799 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.951918 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw2ss\" (UniqueName: \"kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.951980 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952027 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952067 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952179 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952241 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs\") pod \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\" (UID: \"144bf78b-7e67-4e66-a6e9-bf4d2b31be68\") " Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.952909 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.954158 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.961709 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts" (OuterVolumeSpecName: "scripts") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.961688 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss" (OuterVolumeSpecName: "kube-api-access-lw2ss") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "kube-api-access-lw2ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963467 4754 generic.go:334] "Generic (PLEG): container finished" podID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" exitCode=0 Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963519 4754 generic.go:334] "Generic (PLEG): container finished" podID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" exitCode=2 Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963530 4754 generic.go:334] "Generic (PLEG): container finished" podID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" exitCode=0 Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963545 4754 generic.go:334] "Generic (PLEG): container finished" podID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" exitCode=0 Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963578 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerDied","Data":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963626 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerDied","Data":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963640 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerDied","Data":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963643 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963655 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerDied","Data":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963672 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"144bf78b-7e67-4e66-a6e9-bf4d2b31be68","Type":"ContainerDied","Data":"a1c541bf1f4b46c2b2af1d36302b6bbbf0447290f5a792c1003b2ceabb97a114"} Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.963696 4754 scope.go:117] "RemoveContainer" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.990583 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:24 crc kubenswrapper[4754]: I1011 03:57:24.994766 4754 scope.go:117] "RemoveContainer" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.018327 4754 scope.go:117] "RemoveContainer" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.043888 4754 scope.go:117] "RemoveContainer" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.046776 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056109 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056149 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056166 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056181 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056196 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.056208 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw2ss\" (UniqueName: \"kubernetes.io/projected/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-kube-api-access-lw2ss\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.068565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.070878 4754 scope.go:117] "RemoveContainer" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.071488 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": container with ID starting with db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432 not found: ID does not exist" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.071536 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} err="failed to get container status \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": rpc error: code = NotFound desc = could not find container \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": container with ID starting with db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.071575 4754 scope.go:117] "RemoveContainer" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.072337 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": container with ID starting with f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3 not found: ID does not exist" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.072400 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} err="failed to get container status \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": rpc error: code = NotFound desc = could not find container \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": container with ID starting with f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.072451 4754 scope.go:117] "RemoveContainer" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.073032 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": container with ID starting with 20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1 not found: ID does not exist" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.073091 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} err="failed to get container status \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": rpc error: code = NotFound desc = could not find container \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": container with ID starting with 20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.073129 4754 scope.go:117] "RemoveContainer" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.073517 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": container with ID starting with 8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d not found: ID does not exist" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.073591 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} err="failed to get container status \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": rpc error: code = NotFound desc = could not find container \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": container with ID starting with 8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.073608 4754 scope.go:117] "RemoveContainer" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074032 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} err="failed to get container status \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": rpc error: code = NotFound desc = could not find container \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": container with ID starting with db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074062 4754 scope.go:117] "RemoveContainer" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074348 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} err="failed to get container status \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": rpc error: code = NotFound desc = could not find container \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": container with ID starting with f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074373 4754 scope.go:117] "RemoveContainer" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074763 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} err="failed to get container status \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": rpc error: code = NotFound desc = could not find container \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": container with ID starting with 20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.074787 4754 scope.go:117] "RemoveContainer" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075195 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} err="failed to get container status \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": rpc error: code = NotFound desc = could not find container \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": container with ID starting with 8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075221 4754 scope.go:117] "RemoveContainer" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075471 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} err="failed to get container status \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": rpc error: code = NotFound desc = could not find container \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": container with ID starting with db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075498 4754 scope.go:117] "RemoveContainer" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075754 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} err="failed to get container status \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": rpc error: code = NotFound desc = could not find container \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": container with ID starting with f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.075780 4754 scope.go:117] "RemoveContainer" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076291 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} err="failed to get container status \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": rpc error: code = NotFound desc = could not find container \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": container with ID starting with 20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076311 4754 scope.go:117] "RemoveContainer" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076553 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} err="failed to get container status \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": rpc error: code = NotFound desc = could not find container \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": container with ID starting with 8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076588 4754 scope.go:117] "RemoveContainer" containerID="db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076870 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432"} err="failed to get container status \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": rpc error: code = NotFound desc = could not find container \"db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432\": container with ID starting with db7cd28f7d5382df6a552b59a3b5dc2f5c993c654535698ff75e88ad4389f432 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.076896 4754 scope.go:117] "RemoveContainer" containerID="f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.077156 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3"} err="failed to get container status \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": rpc error: code = NotFound desc = could not find container \"f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3\": container with ID starting with f7892c3712347c83a97bf9539b7f561edd60a7412f9b9b526b081f50de599ff3 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.077176 4754 scope.go:117] "RemoveContainer" containerID="20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.077410 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1"} err="failed to get container status \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": rpc error: code = NotFound desc = could not find container \"20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1\": container with ID starting with 20f08dde699e029bc9f770a76356be2af9d25a4a9ea13eebb4c9bcd6fb56c7a1 not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.077448 4754 scope.go:117] "RemoveContainer" containerID="8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.077694 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d"} err="failed to get container status \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": rpc error: code = NotFound desc = could not find container \"8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d\": container with ID starting with 8d09fe2ceea9b2a1ae678a0cec6b260d01bceb47019c4a17de24a0f84514f55d not found: ID does not exist" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.087759 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data" (OuterVolumeSpecName: "config-data") pod "144bf78b-7e67-4e66-a6e9-bf4d2b31be68" (UID: "144bf78b-7e67-4e66-a6e9-bf4d2b31be68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.158479 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.158530 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/144bf78b-7e67-4e66-a6e9-bf4d2b31be68-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.238360 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.308509 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.326165 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.336625 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.337748 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-notification-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.337793 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-notification-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.337900 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-central-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.337918 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-central-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.337957 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="proxy-httpd" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338016 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="proxy-httpd" Oct 11 03:57:25 crc kubenswrapper[4754]: E1011 03:57:25.338058 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="sg-core" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338075 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="sg-core" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338553 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="sg-core" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338593 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="proxy-httpd" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338623 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-notification-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.338664 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" containerName="ceilometer-central-agent" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.344566 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.346606 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.382273 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.382522 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.382918 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480147 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-run-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480253 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-config-data\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480527 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-scripts\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480656 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bz6k\" (UniqueName: \"kubernetes.io/projected/8244086c-5d5e-489a-b798-ed127ec9d89a-kube-api-access-6bz6k\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.480741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-log-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-scripts\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583596 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583619 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bz6k\" (UniqueName: \"kubernetes.io/projected/8244086c-5d5e-489a-b798-ed127ec9d89a-kube-api-access-6bz6k\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-log-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.583946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-run-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.584006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.584044 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-config-data\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.585807 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-run-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.585892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8244086c-5d5e-489a-b798-ed127ec9d89a-log-httpd\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.590005 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-config-data\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.590126 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.591008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.591698 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.603078 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8244086c-5d5e-489a-b798-ed127ec9d89a-scripts\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.607631 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bz6k\" (UniqueName: \"kubernetes.io/projected/8244086c-5d5e-489a-b798-ed127ec9d89a-kube-api-access-6bz6k\") pod \"ceilometer-0\" (UID: \"8244086c-5d5e-489a-b798-ed127ec9d89a\") " pod="openstack/ceilometer-0" Oct 11 03:57:25 crc kubenswrapper[4754]: I1011 03:57:25.694196 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 03:57:26 crc kubenswrapper[4754]: I1011 03:57:26.199201 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 03:57:26 crc kubenswrapper[4754]: I1011 03:57:26.988850 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8244086c-5d5e-489a-b798-ed127ec9d89a","Type":"ContainerStarted","Data":"da507d44d8fe746de8488302a7427e357d57134129dcdd92a82e82739af9293c"} Oct 11 03:57:26 crc kubenswrapper[4754]: I1011 03:57:26.989451 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8244086c-5d5e-489a-b798-ed127ec9d89a","Type":"ContainerStarted","Data":"354062823b0ef46b163fa05f6f8f16b53bd70866dd65073ebf53f58f76ae77b1"} Oct 11 03:57:26 crc kubenswrapper[4754]: I1011 03:57:26.990785 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 11 03:57:27 crc kubenswrapper[4754]: I1011 03:57:27.043543 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:27 crc kubenswrapper[4754]: I1011 03:57:27.096810 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="144bf78b-7e67-4e66-a6e9-bf4d2b31be68" path="/var/lib/kubelet/pods/144bf78b-7e67-4e66-a6e9-bf4d2b31be68/volumes" Oct 11 03:57:28 crc kubenswrapper[4754]: I1011 03:57:28.003227 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8244086c-5d5e-489a-b798-ed127ec9d89a","Type":"ContainerStarted","Data":"4df6900ed808531b895fddfa3474e0cb19e0adf911c097e47a83c6d56c73929c"} Oct 11 03:57:28 crc kubenswrapper[4754]: I1011 03:57:28.003611 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="probe" containerID="cri-o://835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383" gracePeriod=30 Oct 11 03:57:28 crc kubenswrapper[4754]: I1011 03:57:28.003545 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="manila-scheduler" containerID="cri-o://2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211" gracePeriod=30 Oct 11 03:57:29 crc kubenswrapper[4754]: I1011 03:57:29.015007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8244086c-5d5e-489a-b798-ed127ec9d89a","Type":"ContainerStarted","Data":"b6b603597b02c673184707467a6c308045176146fa05355c4cfd2bd5e6ecd28e"} Oct 11 03:57:29 crc kubenswrapper[4754]: I1011 03:57:29.017216 4754 generic.go:334] "Generic (PLEG): container finished" podID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerID="835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383" exitCode=0 Oct 11 03:57:29 crc kubenswrapper[4754]: I1011 03:57:29.017254 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerDied","Data":"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383"} Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.035548 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8244086c-5d5e-489a-b798-ed127ec9d89a","Type":"ContainerStarted","Data":"9605ffff5f05e2cab9b21ba6e7c907af2a0524370c07ddc2d0e5c0d1c3848ad4"} Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.036572 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.087091 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8600493280000001 podStartE2EDuration="5.087067921s" podCreationTimestamp="2025-10-11 03:57:25 +0000 UTC" firstStartedPulling="2025-10-11 03:57:26.207777847 +0000 UTC m=+3093.766722622" lastFinishedPulling="2025-10-11 03:57:29.4347964 +0000 UTC m=+3096.993741215" observedRunningTime="2025-10-11 03:57:30.068641873 +0000 UTC m=+3097.627586698" watchObservedRunningTime="2025-10-11 03:57:30.087067921 +0000 UTC m=+3097.646012706" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.144875 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.239:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.239:8443: connect: connection refused" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.736833 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.736891 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.736935 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.737728 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:57:30 crc kubenswrapper[4754]: I1011 03:57:30.737781 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" gracePeriod=600 Oct 11 03:57:30 crc kubenswrapper[4754]: E1011 03:57:30.865731 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:57:31 crc kubenswrapper[4754]: I1011 03:57:31.052265 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" exitCode=0 Oct 11 03:57:31 crc kubenswrapper[4754]: I1011 03:57:31.052347 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1"} Oct 11 03:57:31 crc kubenswrapper[4754]: I1011 03:57:31.053314 4754 scope.go:117] "RemoveContainer" containerID="fe63dd0c17dfeace13d7f337674186df0807d74c0e332f9436d316a6d291ccd4" Oct 11 03:57:31 crc kubenswrapper[4754]: I1011 03:57:31.055679 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:57:31 crc kubenswrapper[4754]: E1011 03:57:31.058707 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:57:32 crc kubenswrapper[4754]: I1011 03:57:32.378538 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.672744 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.827257 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.827430 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.827491 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.828503 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.828600 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.828592 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.829081 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twzh8\" (UniqueName: \"kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8\") pod \"8d8de385-c514-4fcf-bb9c-6715dbb81931\" (UID: \"8d8de385-c514-4fcf-bb9c-6715dbb81931\") " Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.829926 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d8de385-c514-4fcf-bb9c-6715dbb81931-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.835572 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts" (OuterVolumeSpecName: "scripts") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.837491 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8" (OuterVolumeSpecName: "kube-api-access-twzh8") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "kube-api-access-twzh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.850955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.928251 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.933889 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.933929 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.933943 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.933954 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twzh8\" (UniqueName: \"kubernetes.io/projected/8d8de385-c514-4fcf-bb9c-6715dbb81931-kube-api-access-twzh8\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:33 crc kubenswrapper[4754]: I1011 03:57:33.944153 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data" (OuterVolumeSpecName: "config-data") pod "8d8de385-c514-4fcf-bb9c-6715dbb81931" (UID: "8d8de385-c514-4fcf-bb9c-6715dbb81931"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.036095 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d8de385-c514-4fcf-bb9c-6715dbb81931-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.144652 4754 generic.go:334] "Generic (PLEG): container finished" podID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerID="2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211" exitCode=0 Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.144756 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.144808 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerDied","Data":"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211"} Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.145197 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d8de385-c514-4fcf-bb9c-6715dbb81931","Type":"ContainerDied","Data":"1c894501d8756a6fa67fad90677cac92da1ef3ce2d5d338fcce99fd13663d8f4"} Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.145246 4754 scope.go:117] "RemoveContainer" containerID="835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.175145 4754 scope.go:117] "RemoveContainer" containerID="2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.209300 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.231884 4754 scope.go:117] "RemoveContainer" containerID="835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383" Oct 11 03:57:34 crc kubenswrapper[4754]: E1011 03:57:34.232588 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383\": container with ID starting with 835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383 not found: ID does not exist" containerID="835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.232630 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383"} err="failed to get container status \"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383\": rpc error: code = NotFound desc = could not find container \"835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383\": container with ID starting with 835ae7d36a8182f1981197796e1a9a3553b429b0206c8376a9098daed393e383 not found: ID does not exist" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.232678 4754 scope.go:117] "RemoveContainer" containerID="2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211" Oct 11 03:57:34 crc kubenswrapper[4754]: E1011 03:57:34.232888 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211\": container with ID starting with 2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211 not found: ID does not exist" containerID="2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.232923 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211"} err="failed to get container status \"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211\": rpc error: code = NotFound desc = could not find container \"2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211\": container with ID starting with 2a75bb6613939a71323f14cbe93a608885621c6c16313da7090a82b5aae71211 not found: ID does not exist" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.270178 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.291297 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:34 crc kubenswrapper[4754]: E1011 03:57:34.291888 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="manila-scheduler" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.291911 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="manila-scheduler" Oct 11 03:57:34 crc kubenswrapper[4754]: E1011 03:57:34.291993 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="probe" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.292001 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="probe" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.292225 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="manila-scheduler" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.292252 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" containerName="probe" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.294570 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.300247 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.302220 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.358776 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsfzv\" (UniqueName: \"kubernetes.io/projected/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-kube-api-access-rsfzv\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.358844 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.358901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-scripts\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.358919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.358936 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.359106 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsfzv\" (UniqueName: \"kubernetes.io/projected/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-kube-api-access-rsfzv\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461253 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-scripts\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461401 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.461705 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.467623 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-scripts\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.468545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.471797 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-config-data\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.472063 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.478230 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsfzv\" (UniqueName: \"kubernetes.io/projected/ce406d24-4a3d-4ee8-8793-b61a0a3e6dde-kube-api-access-rsfzv\") pod \"manila-scheduler-0\" (UID: \"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde\") " pod="openstack/manila-scheduler-0" Oct 11 03:57:34 crc kubenswrapper[4754]: I1011 03:57:34.632586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 11 03:57:35 crc kubenswrapper[4754]: I1011 03:57:35.110046 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d8de385-c514-4fcf-bb9c-6715dbb81931" path="/var/lib/kubelet/pods/8d8de385-c514-4fcf-bb9c-6715dbb81931/volumes" Oct 11 03:57:35 crc kubenswrapper[4754]: I1011 03:57:35.145807 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 11 03:57:35 crc kubenswrapper[4754]: W1011 03:57:35.151026 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce406d24_4a3d_4ee8_8793_b61a0a3e6dde.slice/crio-251b9841a2f0f6d7765d5346e7bcbe98e48e16af02f92f165f220b32e70e4025 WatchSource:0}: Error finding container 251b9841a2f0f6d7765d5346e7bcbe98e48e16af02f92f165f220b32e70e4025: Status 404 returned error can't find the container with id 251b9841a2f0f6d7765d5346e7bcbe98e48e16af02f92f165f220b32e70e4025 Oct 11 03:57:36 crc kubenswrapper[4754]: I1011 03:57:36.184656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde","Type":"ContainerStarted","Data":"4c191ec44676b56a9ad569c134dda79b5976c1cca57a33a9cc10524a708d54d3"} Oct 11 03:57:36 crc kubenswrapper[4754]: I1011 03:57:36.185205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde","Type":"ContainerStarted","Data":"6f771997cda3aff3558be0f11f3833712bd10decfe575c4bba2179f510299013"} Oct 11 03:57:36 crc kubenswrapper[4754]: I1011 03:57:36.185217 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ce406d24-4a3d-4ee8-8793-b61a0a3e6dde","Type":"ContainerStarted","Data":"251b9841a2f0f6d7765d5346e7bcbe98e48e16af02f92f165f220b32e70e4025"} Oct 11 03:57:36 crc kubenswrapper[4754]: I1011 03:57:36.221099 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.221075978 podStartE2EDuration="2.221075978s" podCreationTimestamp="2025-10-11 03:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:57:36.21224402 +0000 UTC m=+3103.771188815" watchObservedRunningTime="2025-10-11 03:57:36.221075978 +0000 UTC m=+3103.780020773" Oct 11 03:57:37 crc kubenswrapper[4754]: I1011 03:57:37.017166 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 11 03:57:37 crc kubenswrapper[4754]: I1011 03:57:37.070502 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:37 crc kubenswrapper[4754]: I1011 03:57:37.199877 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="manila-share" containerID="cri-o://644ca25a9d632e868c3f1e02b7cc04817e1d4f77f3814715b396360a240f4de7" gracePeriod=30 Oct 11 03:57:37 crc kubenswrapper[4754]: I1011 03:57:37.200448 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="probe" containerID="cri-o://c68f8833329e3293a0fe38e7b4be65ecd166fb3f8c585e7163e5481a07560123" gracePeriod=30 Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.211609 4754 generic.go:334] "Generic (PLEG): container finished" podID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerID="c68f8833329e3293a0fe38e7b4be65ecd166fb3f8c585e7163e5481a07560123" exitCode=0 Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.212420 4754 generic.go:334] "Generic (PLEG): container finished" podID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerID="644ca25a9d632e868c3f1e02b7cc04817e1d4f77f3814715b396360a240f4de7" exitCode=1 Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.211653 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerDied","Data":"c68f8833329e3293a0fe38e7b4be65ecd166fb3f8c585e7163e5481a07560123"} Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.212457 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerDied","Data":"644ca25a9d632e868c3f1e02b7cc04817e1d4f77f3814715b396360a240f4de7"} Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.321002 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.468752 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.469413 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.469517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.469636 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.469800 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.470410 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqg2t\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.470510 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.470624 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.470748 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts\") pod \"d4405bdb-45d1-4004-b4cf-c766beb8301e\" (UID: \"d4405bdb-45d1-4004-b4cf-c766beb8301e\") " Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.471139 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.471495 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.471521 4754 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d4405bdb-45d1-4004-b4cf-c766beb8301e-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.477950 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts" (OuterVolumeSpecName: "scripts") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.478042 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.478274 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t" (OuterVolumeSpecName: "kube-api-access-mqg2t") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "kube-api-access-mqg2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.479934 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph" (OuterVolumeSpecName: "ceph") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.529868 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.573354 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.573400 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.573413 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.573424 4754 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-ceph\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.573431 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqg2t\" (UniqueName: \"kubernetes.io/projected/d4405bdb-45d1-4004-b4cf-c766beb8301e-kube-api-access-mqg2t\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.587907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data" (OuterVolumeSpecName: "config-data") pod "d4405bdb-45d1-4004-b4cf-c766beb8301e" (UID: "d4405bdb-45d1-4004-b4cf-c766beb8301e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:38 crc kubenswrapper[4754]: I1011 03:57:38.675697 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4405bdb-45d1-4004-b4cf-c766beb8301e-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.226324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d4405bdb-45d1-4004-b4cf-c766beb8301e","Type":"ContainerDied","Data":"f1a2b3f87e8c78a7cade995140ca884a14ee8106ff35252200a64df8c0feeb36"} Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.226384 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.226415 4754 scope.go:117] "RemoveContainer" containerID="c68f8833329e3293a0fe38e7b4be65ecd166fb3f8c585e7163e5481a07560123" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.254560 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.257912 4754 scope.go:117] "RemoveContainer" containerID="644ca25a9d632e868c3f1e02b7cc04817e1d4f77f3814715b396360a240f4de7" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.264638 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.289875 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:39 crc kubenswrapper[4754]: E1011 03:57:39.294108 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="manila-share" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.294134 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="manila-share" Oct 11 03:57:39 crc kubenswrapper[4754]: E1011 03:57:39.294158 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="probe" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.294166 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="probe" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.294468 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="probe" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.294508 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" containerName="manila-share" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.295738 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.298931 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.312336 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.392739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.392807 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.392902 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.392983 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.393170 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-scripts\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.393222 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.393365 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpbrb\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-kube-api-access-dpbrb\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.393621 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-ceph\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpbrb\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-kube-api-access-dpbrb\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-ceph\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496349 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496408 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496483 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-scripts\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496696 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.496833 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/593cef67-51db-4722-a6c9-886fa38cc451-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.503100 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-ceph\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.503130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.503637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.504103 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.507048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/593cef67-51db-4722-a6c9-886fa38cc451-scripts\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.513932 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpbrb\" (UniqueName: \"kubernetes.io/projected/593cef67-51db-4722-a6c9-886fa38cc451-kube-api-access-dpbrb\") pod \"manila-share-share1-0\" (UID: \"593cef67-51db-4722-a6c9-886fa38cc451\") " pod="openstack/manila-share-share1-0" Oct 11 03:57:39 crc kubenswrapper[4754]: I1011 03:57:39.619109 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 11 03:57:40 crc kubenswrapper[4754]: I1011 03:57:40.144770 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-586ffbcfb8-rqrkz" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.239:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.239:8443: connect: connection refused" Oct 11 03:57:40 crc kubenswrapper[4754]: I1011 03:57:40.145504 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:57:40 crc kubenswrapper[4754]: I1011 03:57:40.258919 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 11 03:57:41 crc kubenswrapper[4754]: I1011 03:57:41.099138 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4405bdb-45d1-4004-b4cf-c766beb8301e" path="/var/lib/kubelet/pods/d4405bdb-45d1-4004-b4cf-c766beb8301e/volumes" Oct 11 03:57:41 crc kubenswrapper[4754]: I1011 03:57:41.275262 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"593cef67-51db-4722-a6c9-886fa38cc451","Type":"ContainerStarted","Data":"73a1bf0c2100ba41f411e49a64bd24d201cc4b003a36a0356d27205f09519e63"} Oct 11 03:57:41 crc kubenswrapper[4754]: I1011 03:57:41.275824 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"593cef67-51db-4722-a6c9-886fa38cc451","Type":"ContainerStarted","Data":"5547b99d7feb8bf0cbe0fa8ab05acce60be52c5bd75ea52123b3162ec6123f08"} Oct 11 03:57:42 crc kubenswrapper[4754]: I1011 03:57:42.305537 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"593cef67-51db-4722-a6c9-886fa38cc451","Type":"ContainerStarted","Data":"0a6e1614606eb4575831155a38a4781c4013ef0a5c6779d47e14175a6e09a755"} Oct 11 03:57:42 crc kubenswrapper[4754]: I1011 03:57:42.339721 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.339695262 podStartE2EDuration="3.339695262s" podCreationTimestamp="2025-10-11 03:57:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:57:42.331677677 +0000 UTC m=+3109.890622462" watchObservedRunningTime="2025-10-11 03:57:42.339695262 +0000 UTC m=+3109.898640047" Oct 11 03:57:44 crc kubenswrapper[4754]: I1011 03:57:44.084000 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:57:44 crc kubenswrapper[4754]: E1011 03:57:44.084844 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:57:44 crc kubenswrapper[4754]: I1011 03:57:44.633318 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.152081 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.334248 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.334768 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs" (OuterVolumeSpecName: "logs") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.334994 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjpbq\" (UniqueName: \"kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.335268 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.335646 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.335884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.336272 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.336532 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key\") pod \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\" (UID: \"6d7875e4-f9f8-459d-9b07-cebe349a58b0\") " Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.338304 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d7875e4-f9f8-459d-9b07-cebe349a58b0-logs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.349273 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.349358 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq" (OuterVolumeSpecName: "kube-api-access-sjpbq") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "kube-api-access-sjpbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.375309 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data" (OuterVolumeSpecName: "config-data") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.398691 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts" (OuterVolumeSpecName: "scripts") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.401380 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.403487 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerID="fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121" exitCode=137 Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.403564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerDied","Data":"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121"} Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.403593 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-586ffbcfb8-rqrkz" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.403625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-586ffbcfb8-rqrkz" event={"ID":"6d7875e4-f9f8-459d-9b07-cebe349a58b0","Type":"ContainerDied","Data":"2ab2a5ab20289474552d2b3f07531fb8edc4b405b74e44cb24944fe5bc4132a6"} Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.403659 4754 scope.go:117] "RemoveContainer" containerID="173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.433344 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "6d7875e4-f9f8-459d-9b07-cebe349a58b0" (UID: "6d7875e4-f9f8-459d-9b07-cebe349a58b0"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441806 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441837 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441853 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441866 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d7875e4-f9f8-459d-9b07-cebe349a58b0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441878 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjpbq\" (UniqueName: \"kubernetes.io/projected/6d7875e4-f9f8-459d-9b07-cebe349a58b0-kube-api-access-sjpbq\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.441892 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d7875e4-f9f8-459d-9b07-cebe349a58b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.679517 4754 scope.go:117] "RemoveContainer" containerID="fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.714384 4754 scope.go:117] "RemoveContainer" containerID="173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e" Oct 11 03:57:46 crc kubenswrapper[4754]: E1011 03:57:46.715026 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e\": container with ID starting with 173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e not found: ID does not exist" containerID="173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.715071 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e"} err="failed to get container status \"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e\": rpc error: code = NotFound desc = could not find container \"173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e\": container with ID starting with 173128981a26b7f6708921a8acb785f07ad840450271031a37c727fd6b5a438e not found: ID does not exist" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.715112 4754 scope.go:117] "RemoveContainer" containerID="fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121" Oct 11 03:57:46 crc kubenswrapper[4754]: E1011 03:57:46.715617 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121\": container with ID starting with fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121 not found: ID does not exist" containerID="fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.715703 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121"} err="failed to get container status \"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121\": rpc error: code = NotFound desc = could not find container \"fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121\": container with ID starting with fc460f97f0bc77aab3780b2bd70cd2fb7610c1b1944028ed2630e832cf9c8121 not found: ID does not exist" Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.763321 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:57:46 crc kubenswrapper[4754]: I1011 03:57:46.773175 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-586ffbcfb8-rqrkz"] Oct 11 03:57:47 crc kubenswrapper[4754]: I1011 03:57:47.104537 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" path="/var/lib/kubelet/pods/6d7875e4-f9f8-459d-9b07-cebe349a58b0/volumes" Oct 11 03:57:49 crc kubenswrapper[4754]: I1011 03:57:49.620348 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 11 03:57:55 crc kubenswrapper[4754]: I1011 03:57:55.711560 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 03:57:56 crc kubenswrapper[4754]: I1011 03:57:56.491693 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.806047 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:57:57 crc kubenswrapper[4754]: E1011 03:57:57.806981 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon-log" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.807001 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon-log" Oct 11 03:57:57 crc kubenswrapper[4754]: E1011 03:57:57.807024 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.807032 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.807300 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon-log" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.807330 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7875e4-f9f8-459d-9b07-cebe349a58b0" containerName="horizon" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.808868 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.870022 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.974982 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bg5h\" (UniqueName: \"kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.975082 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:57 crc kubenswrapper[4754]: I1011 03:57:57.975183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.077710 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.077865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bg5h\" (UniqueName: \"kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.078029 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.078853 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.079164 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.113354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bg5h\" (UniqueName: \"kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h\") pod \"certified-operators-zlxkc\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.141724 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:57:58 crc kubenswrapper[4754]: I1011 03:57:58.677920 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:57:58 crc kubenswrapper[4754]: W1011 03:57:58.682531 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod314c35e7_88c4_4cd8_87a7_d585fb7b13a5.slice/crio-f1f48511f01898365c022fcd64dbfe2fe3c3d8a0f7c4f4b438d9ce028aabe4d9 WatchSource:0}: Error finding container f1f48511f01898365c022fcd64dbfe2fe3c3d8a0f7c4f4b438d9ce028aabe4d9: Status 404 returned error can't find the container with id f1f48511f01898365c022fcd64dbfe2fe3c3d8a0f7c4f4b438d9ce028aabe4d9 Oct 11 03:57:59 crc kubenswrapper[4754]: I1011 03:57:59.085845 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:57:59 crc kubenswrapper[4754]: E1011 03:57:59.086651 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:57:59 crc kubenswrapper[4754]: I1011 03:57:59.588956 4754 generic.go:334] "Generic (PLEG): container finished" podID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerID="c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2" exitCode=0 Oct 11 03:57:59 crc kubenswrapper[4754]: I1011 03:57:59.589152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerDied","Data":"c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2"} Oct 11 03:57:59 crc kubenswrapper[4754]: I1011 03:57:59.589797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerStarted","Data":"f1f48511f01898365c022fcd64dbfe2fe3c3d8a0f7c4f4b438d9ce028aabe4d9"} Oct 11 03:58:00 crc kubenswrapper[4754]: I1011 03:58:00.602828 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerStarted","Data":"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832"} Oct 11 03:58:01 crc kubenswrapper[4754]: I1011 03:58:01.209900 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 11 03:58:01 crc kubenswrapper[4754]: I1011 03:58:01.617686 4754 generic.go:334] "Generic (PLEG): container finished" podID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerID="6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832" exitCode=0 Oct 11 03:58:01 crc kubenswrapper[4754]: I1011 03:58:01.617740 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerDied","Data":"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832"} Oct 11 03:58:02 crc kubenswrapper[4754]: I1011 03:58:02.629067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerStarted","Data":"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0"} Oct 11 03:58:02 crc kubenswrapper[4754]: I1011 03:58:02.655545 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zlxkc" podStartSLOduration=3.164465557 podStartE2EDuration="5.655525295s" podCreationTimestamp="2025-10-11 03:57:57 +0000 UTC" firstStartedPulling="2025-10-11 03:57:59.593329767 +0000 UTC m=+3127.152274562" lastFinishedPulling="2025-10-11 03:58:02.084389515 +0000 UTC m=+3129.643334300" observedRunningTime="2025-10-11 03:58:02.647078487 +0000 UTC m=+3130.206023282" watchObservedRunningTime="2025-10-11 03:58:02.655525295 +0000 UTC m=+3130.214470080" Oct 11 03:58:08 crc kubenswrapper[4754]: I1011 03:58:08.142445 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:08 crc kubenswrapper[4754]: I1011 03:58:08.143302 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:08 crc kubenswrapper[4754]: I1011 03:58:08.189180 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:08 crc kubenswrapper[4754]: I1011 03:58:08.748379 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:08 crc kubenswrapper[4754]: I1011 03:58:08.809894 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:58:10 crc kubenswrapper[4754]: I1011 03:58:10.718246 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zlxkc" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="registry-server" containerID="cri-o://cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0" gracePeriod=2 Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.291940 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.418392 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bg5h\" (UniqueName: \"kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h\") pod \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.418492 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content\") pod \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.418594 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities\") pod \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\" (UID: \"314c35e7-88c4-4cd8-87a7-d585fb7b13a5\") " Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.421251 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities" (OuterVolumeSpecName: "utilities") pod "314c35e7-88c4-4cd8-87a7-d585fb7b13a5" (UID: "314c35e7-88c4-4cd8-87a7-d585fb7b13a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.426707 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h" (OuterVolumeSpecName: "kube-api-access-9bg5h") pod "314c35e7-88c4-4cd8-87a7-d585fb7b13a5" (UID: "314c35e7-88c4-4cd8-87a7-d585fb7b13a5"). InnerVolumeSpecName "kube-api-access-9bg5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.470200 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "314c35e7-88c4-4cd8-87a7-d585fb7b13a5" (UID: "314c35e7-88c4-4cd8-87a7-d585fb7b13a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.521174 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bg5h\" (UniqueName: \"kubernetes.io/projected/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-kube-api-access-9bg5h\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.521203 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.521212 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314c35e7-88c4-4cd8-87a7-d585fb7b13a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.734514 4754 generic.go:334] "Generic (PLEG): container finished" podID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerID="cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0" exitCode=0 Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.734593 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerDied","Data":"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0"} Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.734675 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlxkc" event={"ID":"314c35e7-88c4-4cd8-87a7-d585fb7b13a5","Type":"ContainerDied","Data":"f1f48511f01898365c022fcd64dbfe2fe3c3d8a0f7c4f4b438d9ce028aabe4d9"} Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.734692 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlxkc" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.734742 4754 scope.go:117] "RemoveContainer" containerID="cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.772357 4754 scope.go:117] "RemoveContainer" containerID="6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.797634 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.815459 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zlxkc"] Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.817370 4754 scope.go:117] "RemoveContainer" containerID="c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.868328 4754 scope.go:117] "RemoveContainer" containerID="cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0" Oct 11 03:58:11 crc kubenswrapper[4754]: E1011 03:58:11.870160 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0\": container with ID starting with cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0 not found: ID does not exist" containerID="cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.870231 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0"} err="failed to get container status \"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0\": rpc error: code = NotFound desc = could not find container \"cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0\": container with ID starting with cd0f2f28a263b7dbce6a6f1761a9b6debba6c67b14bc4ae805d00b7fb76753c0 not found: ID does not exist" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.870286 4754 scope.go:117] "RemoveContainer" containerID="6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832" Oct 11 03:58:11 crc kubenswrapper[4754]: E1011 03:58:11.870845 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832\": container with ID starting with 6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832 not found: ID does not exist" containerID="6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.870889 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832"} err="failed to get container status \"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832\": rpc error: code = NotFound desc = could not find container \"6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832\": container with ID starting with 6f08148639c08a138586a40ae2e7140d59b35562a03dd1a82e4439e928efc832 not found: ID does not exist" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.870917 4754 scope.go:117] "RemoveContainer" containerID="c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2" Oct 11 03:58:11 crc kubenswrapper[4754]: E1011 03:58:11.871303 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2\": container with ID starting with c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2 not found: ID does not exist" containerID="c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2" Oct 11 03:58:11 crc kubenswrapper[4754]: I1011 03:58:11.871339 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2"} err="failed to get container status \"c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2\": rpc error: code = NotFound desc = could not find container \"c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2\": container with ID starting with c5bd5ef0cf5a38e1d56394a45453ae4a44c47d0b8f74cc7de598b8e6e316c5a2 not found: ID does not exist" Oct 11 03:58:12 crc kubenswrapper[4754]: I1011 03:58:12.084099 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:58:12 crc kubenswrapper[4754]: E1011 03:58:12.084750 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:58:13 crc kubenswrapper[4754]: I1011 03:58:13.099418 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" path="/var/lib/kubelet/pods/314c35e7-88c4-4cd8-87a7-d585fb7b13a5/volumes" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.042685 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:18 crc kubenswrapper[4754]: E1011 03:58:18.044205 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="extract-utilities" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.044227 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="extract-utilities" Oct 11 03:58:18 crc kubenswrapper[4754]: E1011 03:58:18.044250 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="registry-server" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.044257 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="registry-server" Oct 11 03:58:18 crc kubenswrapper[4754]: E1011 03:58:18.044286 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="extract-content" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.044292 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="extract-content" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.044518 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="314c35e7-88c4-4cd8-87a7-d585fb7b13a5" containerName="registry-server" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.047082 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.077802 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.108608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.108938 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjwmp\" (UniqueName: \"kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.109046 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.215199 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjwmp\" (UniqueName: \"kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.215308 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.215551 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.216835 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.217617 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.252646 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjwmp\" (UniqueName: \"kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp\") pod \"redhat-marketplace-gtv95\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.392902 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:18 crc kubenswrapper[4754]: I1011 03:58:18.917707 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:19 crc kubenswrapper[4754]: I1011 03:58:19.845288 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerDied","Data":"d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62"} Oct 11 03:58:19 crc kubenswrapper[4754]: I1011 03:58:19.845862 4754 generic.go:334] "Generic (PLEG): container finished" podID="35f015da-1945-43b8-a342-d311c6cf234f" containerID="d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62" exitCode=0 Oct 11 03:58:19 crc kubenswrapper[4754]: I1011 03:58:19.846217 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerStarted","Data":"e4eab4f3eb60130791c29653d18a973a3af5155346493a79af62815300fcb54a"} Oct 11 03:58:21 crc kubenswrapper[4754]: I1011 03:58:21.906267 4754 generic.go:334] "Generic (PLEG): container finished" podID="35f015da-1945-43b8-a342-d311c6cf234f" containerID="d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3" exitCode=0 Oct 11 03:58:21 crc kubenswrapper[4754]: I1011 03:58:21.906394 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerDied","Data":"d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3"} Oct 11 03:58:22 crc kubenswrapper[4754]: I1011 03:58:22.920795 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerStarted","Data":"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027"} Oct 11 03:58:22 crc kubenswrapper[4754]: I1011 03:58:22.949615 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gtv95" podStartSLOduration=2.350839951 podStartE2EDuration="4.949592798s" podCreationTimestamp="2025-10-11 03:58:18 +0000 UTC" firstStartedPulling="2025-10-11 03:58:19.851060817 +0000 UTC m=+3147.410005602" lastFinishedPulling="2025-10-11 03:58:22.449813674 +0000 UTC m=+3150.008758449" observedRunningTime="2025-10-11 03:58:22.944564436 +0000 UTC m=+3150.503509241" watchObservedRunningTime="2025-10-11 03:58:22.949592798 +0000 UTC m=+3150.508537583" Oct 11 03:58:27 crc kubenswrapper[4754]: I1011 03:58:27.084166 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:58:27 crc kubenswrapper[4754]: E1011 03:58:27.085516 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:58:28 crc kubenswrapper[4754]: I1011 03:58:28.394048 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:28 crc kubenswrapper[4754]: I1011 03:58:28.394151 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:28 crc kubenswrapper[4754]: I1011 03:58:28.511727 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:29 crc kubenswrapper[4754]: I1011 03:58:29.098419 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:30 crc kubenswrapper[4754]: I1011 03:58:30.246690 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.026892 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gtv95" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="registry-server" containerID="cri-o://bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027" gracePeriod=2 Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.525251 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.594384 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities\") pod \"35f015da-1945-43b8-a342-d311c6cf234f\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.594884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjwmp\" (UniqueName: \"kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp\") pod \"35f015da-1945-43b8-a342-d311c6cf234f\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.595120 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content\") pod \"35f015da-1945-43b8-a342-d311c6cf234f\" (UID: \"35f015da-1945-43b8-a342-d311c6cf234f\") " Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.595856 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities" (OuterVolumeSpecName: "utilities") pod "35f015da-1945-43b8-a342-d311c6cf234f" (UID: "35f015da-1945-43b8-a342-d311c6cf234f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.603620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp" (OuterVolumeSpecName: "kube-api-access-pjwmp") pod "35f015da-1945-43b8-a342-d311c6cf234f" (UID: "35f015da-1945-43b8-a342-d311c6cf234f"). InnerVolumeSpecName "kube-api-access-pjwmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.610556 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35f015da-1945-43b8-a342-d311c6cf234f" (UID: "35f015da-1945-43b8-a342-d311c6cf234f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.698723 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.698763 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjwmp\" (UniqueName: \"kubernetes.io/projected/35f015da-1945-43b8-a342-d311c6cf234f-kube-api-access-pjwmp\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:31 crc kubenswrapper[4754]: I1011 03:58:31.698780 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35f015da-1945-43b8-a342-d311c6cf234f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.043696 4754 generic.go:334] "Generic (PLEG): container finished" podID="35f015da-1945-43b8-a342-d311c6cf234f" containerID="bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027" exitCode=0 Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.043778 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerDied","Data":"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027"} Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.043839 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtv95" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.043872 4754 scope.go:117] "RemoveContainer" containerID="bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.043848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtv95" event={"ID":"35f015da-1945-43b8-a342-d311c6cf234f","Type":"ContainerDied","Data":"e4eab4f3eb60130791c29653d18a973a3af5155346493a79af62815300fcb54a"} Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.074980 4754 scope.go:117] "RemoveContainer" containerID="d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.104444 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.113736 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtv95"] Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.123648 4754 scope.go:117] "RemoveContainer" containerID="d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.162373 4754 scope.go:117] "RemoveContainer" containerID="bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027" Oct 11 03:58:32 crc kubenswrapper[4754]: E1011 03:58:32.163626 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027\": container with ID starting with bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027 not found: ID does not exist" containerID="bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.163695 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027"} err="failed to get container status \"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027\": rpc error: code = NotFound desc = could not find container \"bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027\": container with ID starting with bd5e20f4f3ba8c8ea4bba85e371bc6e29d14df3a14dfae795ebdf92a74f10027 not found: ID does not exist" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.163737 4754 scope.go:117] "RemoveContainer" containerID="d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3" Oct 11 03:58:32 crc kubenswrapper[4754]: E1011 03:58:32.164404 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3\": container with ID starting with d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3 not found: ID does not exist" containerID="d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.164585 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3"} err="failed to get container status \"d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3\": rpc error: code = NotFound desc = could not find container \"d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3\": container with ID starting with d2cb4f4dac5cc4fa6b449594a04950fda661e26fbaf7ca0a95ee6636b9b3d4a3 not found: ID does not exist" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.164679 4754 scope.go:117] "RemoveContainer" containerID="d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62" Oct 11 03:58:32 crc kubenswrapper[4754]: E1011 03:58:32.165225 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62\": container with ID starting with d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62 not found: ID does not exist" containerID="d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62" Oct 11 03:58:32 crc kubenswrapper[4754]: I1011 03:58:32.165284 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62"} err="failed to get container status \"d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62\": rpc error: code = NotFound desc = could not find container \"d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62\": container with ID starting with d7c3d6376091481d0b1be0a3072302b5204ecafed6e8912c932c937ade245e62 not found: ID does not exist" Oct 11 03:58:33 crc kubenswrapper[4754]: I1011 03:58:33.121757 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35f015da-1945-43b8-a342-d311c6cf234f" path="/var/lib/kubelet/pods/35f015da-1945-43b8-a342-d311c6cf234f/volumes" Oct 11 03:58:38 crc kubenswrapper[4754]: I1011 03:58:38.084502 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:58:38 crc kubenswrapper[4754]: E1011 03:58:38.085375 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:58:50 crc kubenswrapper[4754]: I1011 03:58:50.084443 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:58:50 crc kubenswrapper[4754]: E1011 03:58:50.085852 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.100548 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 11 03:59:01 crc kubenswrapper[4754]: E1011 03:59:01.102846 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="registry-server" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.102913 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="registry-server" Oct 11 03:59:01 crc kubenswrapper[4754]: E1011 03:59:01.103018 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="extract-utilities" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.103035 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="extract-utilities" Oct 11 03:59:01 crc kubenswrapper[4754]: E1011 03:59:01.103108 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="extract-content" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.103124 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="extract-content" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.103927 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f015da-1945-43b8-a342-d311c6cf234f" containerName="registry-server" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.105823 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.111226 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.116462 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.116497 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.116678 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdw6z" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.116462 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.287454 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.287744 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n2b5\" (UniqueName: \"kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.287799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.287845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.287906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.288111 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.288163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.288280 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.288334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.402936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403048 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403121 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403281 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403417 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403518 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n2b5\" (UniqueName: \"kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403676 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403791 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.403923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.404354 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.405168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.406113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.411288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.413745 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.419649 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.421556 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n2b5\" (UniqueName: \"kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.440782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " pod="openstack/tempest-tests-tempest" Oct 11 03:59:01 crc kubenswrapper[4754]: I1011 03:59:01.744224 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 11 03:59:02 crc kubenswrapper[4754]: I1011 03:59:02.251951 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 11 03:59:02 crc kubenswrapper[4754]: I1011 03:59:02.406115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1b1a049c-cb14-4ec6-8843-69950502bce7","Type":"ContainerStarted","Data":"4ebc0e884da73328a56be46c539189af12ccb1f7325c7ada38130d3ac42764eb"} Oct 11 03:59:05 crc kubenswrapper[4754]: I1011 03:59:05.084956 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:59:05 crc kubenswrapper[4754]: E1011 03:59:05.085701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:59:19 crc kubenswrapper[4754]: I1011 03:59:19.084647 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:59:19 crc kubenswrapper[4754]: E1011 03:59:19.085943 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:59:30 crc kubenswrapper[4754]: E1011 03:59:30.151661 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 11 03:59:30 crc kubenswrapper[4754]: E1011 03:59:30.152725 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7n2b5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1b1a049c-cb14-4ec6-8843-69950502bce7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 03:59:30 crc kubenswrapper[4754]: E1011 03:59:30.154740 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1b1a049c-cb14-4ec6-8843-69950502bce7" Oct 11 03:59:30 crc kubenswrapper[4754]: E1011 03:59:30.713767 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1b1a049c-cb14-4ec6-8843-69950502bce7" Oct 11 03:59:31 crc kubenswrapper[4754]: I1011 03:59:31.083874 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:59:31 crc kubenswrapper[4754]: E1011 03:59:31.084616 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:59:46 crc kubenswrapper[4754]: I1011 03:59:46.084218 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 03:59:46 crc kubenswrapper[4754]: E1011 03:59:46.085414 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 03:59:47 crc kubenswrapper[4754]: I1011 03:59:47.914774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1b1a049c-cb14-4ec6-8843-69950502bce7","Type":"ContainerStarted","Data":"e0f90349f5516615b638747af33a2d943849a0bc18247d391e40f0802afb7754"} Oct 11 03:59:47 crc kubenswrapper[4754]: I1011 03:59:47.943551 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.662710178 podStartE2EDuration="47.9435309s" podCreationTimestamp="2025-10-11 03:59:00 +0000 UTC" firstStartedPulling="2025-10-11 03:59:02.262481554 +0000 UTC m=+3189.821426339" lastFinishedPulling="2025-10-11 03:59:46.543302266 +0000 UTC m=+3234.102247061" observedRunningTime="2025-10-11 03:59:47.936024769 +0000 UTC m=+3235.494969554" watchObservedRunningTime="2025-10-11 03:59:47.9435309 +0000 UTC m=+3235.502475685" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.083871 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:00:00 crc kubenswrapper[4754]: E1011 04:00:00.085140 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.196107 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb"] Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.197869 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.201815 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb"] Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.202493 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.202705 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.392152 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.392902 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtnx\" (UniqueName: \"kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.393212 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.495207 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.495340 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtnx\" (UniqueName: \"kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.495392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.496557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.504782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.526688 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtnx\" (UniqueName: \"kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx\") pod \"collect-profiles-29335920-c54nb\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:00 crc kubenswrapper[4754]: I1011 04:00:00.534599 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:01 crc kubenswrapper[4754]: W1011 04:00:01.028993 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod273fccc7_a208_4fca_990c_ecc01d998324.slice/crio-f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008 WatchSource:0}: Error finding container f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008: Status 404 returned error can't find the container with id f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008 Oct 11 04:00:01 crc kubenswrapper[4754]: I1011 04:00:01.029312 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb"] Oct 11 04:00:01 crc kubenswrapper[4754]: I1011 04:00:01.054304 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" event={"ID":"273fccc7-a208-4fca-990c-ecc01d998324","Type":"ContainerStarted","Data":"f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008"} Oct 11 04:00:02 crc kubenswrapper[4754]: I1011 04:00:02.075672 4754 generic.go:334] "Generic (PLEG): container finished" podID="273fccc7-a208-4fca-990c-ecc01d998324" containerID="e3d95d8a2dd41a6d7a27afacfa8254928190f0e084fed4f301e5b1333a9cf23b" exitCode=0 Oct 11 04:00:02 crc kubenswrapper[4754]: I1011 04:00:02.075784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" event={"ID":"273fccc7-a208-4fca-990c-ecc01d998324","Type":"ContainerDied","Data":"e3d95d8a2dd41a6d7a27afacfa8254928190f0e084fed4f301e5b1333a9cf23b"} Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.474708 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.566017 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume\") pod \"273fccc7-a208-4fca-990c-ecc01d998324\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.566112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume\") pod \"273fccc7-a208-4fca-990c-ecc01d998324\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.566226 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtnx\" (UniqueName: \"kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx\") pod \"273fccc7-a208-4fca-990c-ecc01d998324\" (UID: \"273fccc7-a208-4fca-990c-ecc01d998324\") " Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.567300 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume" (OuterVolumeSpecName: "config-volume") pod "273fccc7-a208-4fca-990c-ecc01d998324" (UID: "273fccc7-a208-4fca-990c-ecc01d998324"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.574344 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx" (OuterVolumeSpecName: "kube-api-access-vrtnx") pod "273fccc7-a208-4fca-990c-ecc01d998324" (UID: "273fccc7-a208-4fca-990c-ecc01d998324"). InnerVolumeSpecName "kube-api-access-vrtnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.576156 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "273fccc7-a208-4fca-990c-ecc01d998324" (UID: "273fccc7-a208-4fca-990c-ecc01d998324"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.669206 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtnx\" (UniqueName: \"kubernetes.io/projected/273fccc7-a208-4fca-990c-ecc01d998324-kube-api-access-vrtnx\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.669264 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273fccc7-a208-4fca-990c-ecc01d998324-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:03 crc kubenswrapper[4754]: I1011 04:00:03.669322 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273fccc7-a208-4fca-990c-ecc01d998324-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:04 crc kubenswrapper[4754]: I1011 04:00:04.098775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" event={"ID":"273fccc7-a208-4fca-990c-ecc01d998324","Type":"ContainerDied","Data":"f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008"} Oct 11 04:00:04 crc kubenswrapper[4754]: I1011 04:00:04.098826 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f319a955c3abc15c49fee61c8f820a119d44b35105e73f83aff513998f81d008" Oct 11 04:00:04 crc kubenswrapper[4754]: I1011 04:00:04.099158 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-c54nb" Oct 11 04:00:04 crc kubenswrapper[4754]: I1011 04:00:04.566517 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq"] Oct 11 04:00:04 crc kubenswrapper[4754]: I1011 04:00:04.575037 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335875-qxtrq"] Oct 11 04:00:05 crc kubenswrapper[4754]: I1011 04:00:05.097826 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9bc4270-b249-44eb-9931-b026cf1721f0" path="/var/lib/kubelet/pods/e9bc4270-b249-44eb-9931-b026cf1721f0/volumes" Oct 11 04:00:06 crc kubenswrapper[4754]: I1011 04:00:06.923656 4754 scope.go:117] "RemoveContainer" containerID="19ba11560fc0ad36ba3b52afbb484b0afc1a78b78c599b27fc84fd47db75a773" Oct 11 04:00:15 crc kubenswrapper[4754]: I1011 04:00:15.084485 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:00:15 crc kubenswrapper[4754]: E1011 04:00:15.086070 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:00:26 crc kubenswrapper[4754]: I1011 04:00:26.086244 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:00:26 crc kubenswrapper[4754]: E1011 04:00:26.088384 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:00:40 crc kubenswrapper[4754]: I1011 04:00:40.084283 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:00:40 crc kubenswrapper[4754]: E1011 04:00:40.085287 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:00:54 crc kubenswrapper[4754]: I1011 04:00:54.084914 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:00:54 crc kubenswrapper[4754]: E1011 04:00:54.086434 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.158794 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29335921-n6ljl"] Oct 11 04:01:00 crc kubenswrapper[4754]: E1011 04:01:00.160078 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="273fccc7-a208-4fca-990c-ecc01d998324" containerName="collect-profiles" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.160099 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="273fccc7-a208-4fca-990c-ecc01d998324" containerName="collect-profiles" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.160415 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="273fccc7-a208-4fca-990c-ecc01d998324" containerName="collect-profiles" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.161474 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.205210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335921-n6ljl"] Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.262470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.262531 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcpsj\" (UniqueName: \"kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.262788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.263439 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.366042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.366356 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.366398 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcpsj\" (UniqueName: \"kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.366577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.377137 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.380896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.385608 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.398688 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcpsj\" (UniqueName: \"kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj\") pod \"keystone-cron-29335921-n6ljl\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:00 crc kubenswrapper[4754]: I1011 04:01:00.495070 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:01 crc kubenswrapper[4754]: I1011 04:01:01.106747 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335921-n6ljl"] Oct 11 04:01:01 crc kubenswrapper[4754]: I1011 04:01:01.773226 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335921-n6ljl" event={"ID":"7d221e22-af78-451c-ae16-7650053c7b4d","Type":"ContainerStarted","Data":"b12fd95df29c3aa5bb053d68a60d137591359d073dc008d60a078c98df7023eb"} Oct 11 04:01:01 crc kubenswrapper[4754]: I1011 04:01:01.775329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335921-n6ljl" event={"ID":"7d221e22-af78-451c-ae16-7650053c7b4d","Type":"ContainerStarted","Data":"a1f4b2ad200aa05f8041cb44431748ee9164697723f4c196fb8c891aeecf6d85"} Oct 11 04:01:01 crc kubenswrapper[4754]: I1011 04:01:01.798912 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29335921-n6ljl" podStartSLOduration=1.798890529 podStartE2EDuration="1.798890529s" podCreationTimestamp="2025-10-11 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:01:01.797165981 +0000 UTC m=+3309.356110776" watchObservedRunningTime="2025-10-11 04:01:01.798890529 +0000 UTC m=+3309.357835304" Oct 11 04:01:03 crc kubenswrapper[4754]: I1011 04:01:03.803482 4754 generic.go:334] "Generic (PLEG): container finished" podID="7d221e22-af78-451c-ae16-7650053c7b4d" containerID="b12fd95df29c3aa5bb053d68a60d137591359d073dc008d60a078c98df7023eb" exitCode=0 Oct 11 04:01:03 crc kubenswrapper[4754]: I1011 04:01:03.803556 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335921-n6ljl" event={"ID":"7d221e22-af78-451c-ae16-7650053c7b4d","Type":"ContainerDied","Data":"b12fd95df29c3aa5bb053d68a60d137591359d073dc008d60a078c98df7023eb"} Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.248369 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.323077 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcpsj\" (UniqueName: \"kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj\") pod \"7d221e22-af78-451c-ae16-7650053c7b4d\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.323429 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data\") pod \"7d221e22-af78-451c-ae16-7650053c7b4d\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.323474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys\") pod \"7d221e22-af78-451c-ae16-7650053c7b4d\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.323550 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle\") pod \"7d221e22-af78-451c-ae16-7650053c7b4d\" (UID: \"7d221e22-af78-451c-ae16-7650053c7b4d\") " Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.331628 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj" (OuterVolumeSpecName: "kube-api-access-dcpsj") pod "7d221e22-af78-451c-ae16-7650053c7b4d" (UID: "7d221e22-af78-451c-ae16-7650053c7b4d"). InnerVolumeSpecName "kube-api-access-dcpsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.332481 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7d221e22-af78-451c-ae16-7650053c7b4d" (UID: "7d221e22-af78-451c-ae16-7650053c7b4d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.361565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d221e22-af78-451c-ae16-7650053c7b4d" (UID: "7d221e22-af78-451c-ae16-7650053c7b4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.402368 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data" (OuterVolumeSpecName: "config-data") pod "7d221e22-af78-451c-ae16-7650053c7b4d" (UID: "7d221e22-af78-451c-ae16-7650053c7b4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.425815 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcpsj\" (UniqueName: \"kubernetes.io/projected/7d221e22-af78-451c-ae16-7650053c7b4d-kube-api-access-dcpsj\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.425856 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.425871 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.425889 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d221e22-af78-451c-ae16-7650053c7b4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.828789 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335921-n6ljl" event={"ID":"7d221e22-af78-451c-ae16-7650053c7b4d","Type":"ContainerDied","Data":"a1f4b2ad200aa05f8041cb44431748ee9164697723f4c196fb8c891aeecf6d85"} Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.828832 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1f4b2ad200aa05f8041cb44431748ee9164697723f4c196fb8c891aeecf6d85" Oct 11 04:01:05 crc kubenswrapper[4754]: I1011 04:01:05.828832 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335921-n6ljl" Oct 11 04:01:09 crc kubenswrapper[4754]: I1011 04:01:09.084030 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:01:09 crc kubenswrapper[4754]: E1011 04:01:09.084852 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:01:20 crc kubenswrapper[4754]: I1011 04:01:20.084399 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:01:20 crc kubenswrapper[4754]: E1011 04:01:20.085701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:01:34 crc kubenswrapper[4754]: I1011 04:01:34.084361 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:01:34 crc kubenswrapper[4754]: E1011 04:01:34.085324 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:01:46 crc kubenswrapper[4754]: I1011 04:01:46.084301 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:01:46 crc kubenswrapper[4754]: E1011 04:01:46.085701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:01:58 crc kubenswrapper[4754]: I1011 04:01:58.084335 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:01:58 crc kubenswrapper[4754]: E1011 04:01:58.085433 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:02:11 crc kubenswrapper[4754]: I1011 04:02:11.083997 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:02:11 crc kubenswrapper[4754]: E1011 04:02:11.086557 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:02:26 crc kubenswrapper[4754]: I1011 04:02:26.084593 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:02:26 crc kubenswrapper[4754]: E1011 04:02:26.085677 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:02:40 crc kubenswrapper[4754]: I1011 04:02:40.085084 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:02:40 crc kubenswrapper[4754]: I1011 04:02:40.910364 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6"} Oct 11 04:03:07 crc kubenswrapper[4754]: I1011 04:03:07.077345 4754 scope.go:117] "RemoveContainer" containerID="6181c8f3937f5a6c3d5e8285ccd45c91d5273e5df3c8cd3eee89a4de7035448f" Oct 11 04:03:07 crc kubenswrapper[4754]: I1011 04:03:07.112766 4754 scope.go:117] "RemoveContainer" containerID="f02ad7f0be515f242c2e6a68855c00dfd21770773115f4e650a475a55312480d" Oct 11 04:05:00 crc kubenswrapper[4754]: I1011 04:05:00.737196 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:05:00 crc kubenswrapper[4754]: I1011 04:05:00.738279 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:05:30 crc kubenswrapper[4754]: I1011 04:05:30.736643 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:05:30 crc kubenswrapper[4754]: I1011 04:05:30.737586 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:06:00 crc kubenswrapper[4754]: I1011 04:06:00.736443 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:06:00 crc kubenswrapper[4754]: I1011 04:06:00.737125 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:06:00 crc kubenswrapper[4754]: I1011 04:06:00.737194 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:06:00 crc kubenswrapper[4754]: I1011 04:06:00.738452 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:06:00 crc kubenswrapper[4754]: I1011 04:06:00.738542 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6" gracePeriod=600 Oct 11 04:06:01 crc kubenswrapper[4754]: I1011 04:06:01.061356 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6" exitCode=0 Oct 11 04:06:01 crc kubenswrapper[4754]: I1011 04:06:01.061449 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6"} Oct 11 04:06:01 crc kubenswrapper[4754]: I1011 04:06:01.061812 4754 scope.go:117] "RemoveContainer" containerID="0fee395251b1c724cf8e32fc6ee7ad7fe57e7ab4d4f17dd546f038e817efccf1" Oct 11 04:06:02 crc kubenswrapper[4754]: I1011 04:06:02.076165 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710"} Oct 11 04:06:31 crc kubenswrapper[4754]: I1011 04:06:31.051413 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-t9ct9"] Oct 11 04:06:31 crc kubenswrapper[4754]: I1011 04:06:31.061513 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-t9ct9"] Oct 11 04:06:31 crc kubenswrapper[4754]: I1011 04:06:31.094147 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10b2f10c-b480-4250-b44c-883892870d1b" path="/var/lib/kubelet/pods/10b2f10c-b480-4250-b44c-883892870d1b/volumes" Oct 11 04:06:43 crc kubenswrapper[4754]: I1011 04:06:43.035273 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-ed00-account-create-q2rgc"] Oct 11 04:06:43 crc kubenswrapper[4754]: I1011 04:06:43.051337 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-ed00-account-create-q2rgc"] Oct 11 04:06:43 crc kubenswrapper[4754]: I1011 04:06:43.096556 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7055e1d-ffb5-4cca-9a53-2400675059a0" path="/var/lib/kubelet/pods/f7055e1d-ffb5-4cca-9a53-2400675059a0/volumes" Oct 11 04:07:04 crc kubenswrapper[4754]: I1011 04:07:04.042615 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-kpkwx"] Oct 11 04:07:04 crc kubenswrapper[4754]: I1011 04:07:04.056289 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-kpkwx"] Oct 11 04:07:05 crc kubenswrapper[4754]: I1011 04:07:05.098992 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2dd8cc-cbb0-4675-b9a4-46846b783730" path="/var/lib/kubelet/pods/5e2dd8cc-cbb0-4675-b9a4-46846b783730/volumes" Oct 11 04:07:07 crc kubenswrapper[4754]: I1011 04:07:07.248837 4754 scope.go:117] "RemoveContainer" containerID="f615f38466605f507c3cc08ba551cc10f5513934ae868e731628053eee2d1c3c" Oct 11 04:07:07 crc kubenswrapper[4754]: I1011 04:07:07.295802 4754 scope.go:117] "RemoveContainer" containerID="d9b934ea544b50dbcffeba14c6bbe95f45b7f0c3562e7cb45b752331ea5e7da7" Oct 11 04:07:07 crc kubenswrapper[4754]: I1011 04:07:07.387578 4754 scope.go:117] "RemoveContainer" containerID="538393f387053279ccb6b2298bda16aec0b28fb06233a7441bedfdc9f57a6fdf" Oct 11 04:08:30 crc kubenswrapper[4754]: I1011 04:08:30.736314 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:08:30 crc kubenswrapper[4754]: I1011 04:08:30.737006 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.677082 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:32 crc kubenswrapper[4754]: E1011 04:08:32.678632 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d221e22-af78-451c-ae16-7650053c7b4d" containerName="keystone-cron" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.678654 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d221e22-af78-451c-ae16-7650053c7b4d" containerName="keystone-cron" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.679189 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d221e22-af78-451c-ae16-7650053c7b4d" containerName="keystone-cron" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.681155 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.697611 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.755802 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.755889 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.756080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nc6f\" (UniqueName: \"kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.858790 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.858889 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nc6f\" (UniqueName: \"kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.859093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.859335 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.859742 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:32 crc kubenswrapper[4754]: I1011 04:08:32.883559 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nc6f\" (UniqueName: \"kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f\") pod \"certified-operators-cb2rc\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:33 crc kubenswrapper[4754]: I1011 04:08:33.019321 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:33 crc kubenswrapper[4754]: I1011 04:08:33.561269 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:33 crc kubenswrapper[4754]: I1011 04:08:33.607378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerStarted","Data":"6db7f9f8ee66c182c8067d6ff3b9b0c28e0104fd7d78532b40b08fdaa00be8bf"} Oct 11 04:08:34 crc kubenswrapper[4754]: I1011 04:08:34.621768 4754 generic.go:334] "Generic (PLEG): container finished" podID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerID="a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c" exitCode=0 Oct 11 04:08:34 crc kubenswrapper[4754]: I1011 04:08:34.621866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerDied","Data":"a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c"} Oct 11 04:08:34 crc kubenswrapper[4754]: I1011 04:08:34.626576 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:08:35 crc kubenswrapper[4754]: I1011 04:08:35.635527 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerStarted","Data":"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00"} Oct 11 04:08:36 crc kubenswrapper[4754]: I1011 04:08:36.647672 4754 generic.go:334] "Generic (PLEG): container finished" podID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerID="fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00" exitCode=0 Oct 11 04:08:36 crc kubenswrapper[4754]: I1011 04:08:36.647765 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerDied","Data":"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00"} Oct 11 04:08:37 crc kubenswrapper[4754]: I1011 04:08:37.661786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerStarted","Data":"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817"} Oct 11 04:08:37 crc kubenswrapper[4754]: I1011 04:08:37.685649 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cb2rc" podStartSLOduration=3.232142844 podStartE2EDuration="5.685625254s" podCreationTimestamp="2025-10-11 04:08:32 +0000 UTC" firstStartedPulling="2025-10-11 04:08:34.624610199 +0000 UTC m=+3762.183555034" lastFinishedPulling="2025-10-11 04:08:37.078092629 +0000 UTC m=+3764.637037444" observedRunningTime="2025-10-11 04:08:37.68084102 +0000 UTC m=+3765.239785815" watchObservedRunningTime="2025-10-11 04:08:37.685625254 +0000 UTC m=+3765.244570039" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.062273 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.065600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.091155 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.171934 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.172112 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rtb5\" (UniqueName: \"kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.172839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.274825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.274914 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rtb5\" (UniqueName: \"kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.275072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.275431 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.275497 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.300452 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rtb5\" (UniqueName: \"kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5\") pod \"redhat-marketplace-6hh5s\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.391679 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:40 crc kubenswrapper[4754]: I1011 04:08:40.891055 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:40 crc kubenswrapper[4754]: W1011 04:08:40.910850 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea70f27f_0c50_41b6_a711_435eb087193a.slice/crio-a758826dfd5f624de2c8c2d385bae4e5588c5fefb61b12da375acab68645acfd WatchSource:0}: Error finding container a758826dfd5f624de2c8c2d385bae4e5588c5fefb61b12da375acab68645acfd: Status 404 returned error can't find the container with id a758826dfd5f624de2c8c2d385bae4e5588c5fefb61b12da375acab68645acfd Oct 11 04:08:41 crc kubenswrapper[4754]: I1011 04:08:41.696496 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea70f27f-0c50-41b6-a711-435eb087193a" containerID="55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c" exitCode=0 Oct 11 04:08:41 crc kubenswrapper[4754]: I1011 04:08:41.696881 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerDied","Data":"55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c"} Oct 11 04:08:41 crc kubenswrapper[4754]: I1011 04:08:41.696917 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerStarted","Data":"a758826dfd5f624de2c8c2d385bae4e5588c5fefb61b12da375acab68645acfd"} Oct 11 04:08:42 crc kubenswrapper[4754]: I1011 04:08:42.710579 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea70f27f-0c50-41b6-a711-435eb087193a" containerID="0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32" exitCode=0 Oct 11 04:08:42 crc kubenswrapper[4754]: I1011 04:08:42.710814 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerDied","Data":"0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32"} Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.019560 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.019636 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.081621 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.724376 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerStarted","Data":"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc"} Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.749031 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6hh5s" podStartSLOduration=2.074985769 podStartE2EDuration="3.749005083s" podCreationTimestamp="2025-10-11 04:08:40 +0000 UTC" firstStartedPulling="2025-10-11 04:08:41.698937189 +0000 UTC m=+3769.257881984" lastFinishedPulling="2025-10-11 04:08:43.372956513 +0000 UTC m=+3770.931901298" observedRunningTime="2025-10-11 04:08:43.746629537 +0000 UTC m=+3771.305574332" watchObservedRunningTime="2025-10-11 04:08:43.749005083 +0000 UTC m=+3771.307949878" Oct 11 04:08:43 crc kubenswrapper[4754]: I1011 04:08:43.801391 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:45 crc kubenswrapper[4754]: I1011 04:08:45.444834 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:45 crc kubenswrapper[4754]: I1011 04:08:45.749942 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cb2rc" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="registry-server" containerID="cri-o://844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817" gracePeriod=2 Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.438633 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.564695 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nc6f\" (UniqueName: \"kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f\") pod \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.564875 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content\") pod \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.564953 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities\") pod \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\" (UID: \"eb8cab65-7c24-4c8b-9e60-4134f080c17a\") " Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.565897 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities" (OuterVolumeSpecName: "utilities") pod "eb8cab65-7c24-4c8b-9e60-4134f080c17a" (UID: "eb8cab65-7c24-4c8b-9e60-4134f080c17a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.576670 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f" (OuterVolumeSpecName: "kube-api-access-8nc6f") pod "eb8cab65-7c24-4c8b-9e60-4134f080c17a" (UID: "eb8cab65-7c24-4c8b-9e60-4134f080c17a"). InnerVolumeSpecName "kube-api-access-8nc6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.630450 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb8cab65-7c24-4c8b-9e60-4134f080c17a" (UID: "eb8cab65-7c24-4c8b-9e60-4134f080c17a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.668305 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nc6f\" (UniqueName: \"kubernetes.io/projected/eb8cab65-7c24-4c8b-9e60-4134f080c17a-kube-api-access-8nc6f\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.668342 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.668352 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8cab65-7c24-4c8b-9e60-4134f080c17a-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.763828 4754 generic.go:334] "Generic (PLEG): container finished" podID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerID="844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817" exitCode=0 Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.763929 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerDied","Data":"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817"} Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.763984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2rc" event={"ID":"eb8cab65-7c24-4c8b-9e60-4134f080c17a","Type":"ContainerDied","Data":"6db7f9f8ee66c182c8067d6ff3b9b0c28e0104fd7d78532b40b08fdaa00be8bf"} Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.764010 4754 scope.go:117] "RemoveContainer" containerID="844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.764208 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2rc" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.798098 4754 scope.go:117] "RemoveContainer" containerID="fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.800926 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.809014 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cb2rc"] Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.841822 4754 scope.go:117] "RemoveContainer" containerID="a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.878704 4754 scope.go:117] "RemoveContainer" containerID="844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817" Oct 11 04:08:46 crc kubenswrapper[4754]: E1011 04:08:46.880918 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817\": container with ID starting with 844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817 not found: ID does not exist" containerID="844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.880984 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817"} err="failed to get container status \"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817\": rpc error: code = NotFound desc = could not find container \"844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817\": container with ID starting with 844e271925798835e8b7055506345bd35fc1a2cfe78aef2f111b1a9c7dcde817 not found: ID does not exist" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.881026 4754 scope.go:117] "RemoveContainer" containerID="fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00" Oct 11 04:08:46 crc kubenswrapper[4754]: E1011 04:08:46.882383 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00\": container with ID starting with fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00 not found: ID does not exist" containerID="fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.882432 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00"} err="failed to get container status \"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00\": rpc error: code = NotFound desc = could not find container \"fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00\": container with ID starting with fd9d694b97d09784b0bbc9b5aca2f2aa4f97de0fa6746685d21777b363a43b00 not found: ID does not exist" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.882468 4754 scope.go:117] "RemoveContainer" containerID="a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c" Oct 11 04:08:46 crc kubenswrapper[4754]: E1011 04:08:46.887441 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c\": container with ID starting with a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c not found: ID does not exist" containerID="a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c" Oct 11 04:08:46 crc kubenswrapper[4754]: I1011 04:08:46.887564 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c"} err="failed to get container status \"a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c\": rpc error: code = NotFound desc = could not find container \"a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c\": container with ID starting with a99dd48763fe3e7f19f6eeb04ad302db15cf2c73ae82958bf9f6f6d06bc84b0c not found: ID does not exist" Oct 11 04:08:47 crc kubenswrapper[4754]: I1011 04:08:47.095163 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" path="/var/lib/kubelet/pods/eb8cab65-7c24-4c8b-9e60-4134f080c17a/volumes" Oct 11 04:08:50 crc kubenswrapper[4754]: I1011 04:08:50.392565 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:50 crc kubenswrapper[4754]: I1011 04:08:50.393024 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:50 crc kubenswrapper[4754]: I1011 04:08:50.506621 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:51 crc kubenswrapper[4754]: I1011 04:08:51.226284 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:51 crc kubenswrapper[4754]: I1011 04:08:51.347261 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:52 crc kubenswrapper[4754]: I1011 04:08:52.821058 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6hh5s" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="registry-server" containerID="cri-o://81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc" gracePeriod=2 Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.595887 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.761280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rtb5\" (UniqueName: \"kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5\") pod \"ea70f27f-0c50-41b6-a711-435eb087193a\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.761382 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities\") pod \"ea70f27f-0c50-41b6-a711-435eb087193a\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.761517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content\") pod \"ea70f27f-0c50-41b6-a711-435eb087193a\" (UID: \"ea70f27f-0c50-41b6-a711-435eb087193a\") " Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.762460 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities" (OuterVolumeSpecName: "utilities") pod "ea70f27f-0c50-41b6-a711-435eb087193a" (UID: "ea70f27f-0c50-41b6-a711-435eb087193a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.762861 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.776499 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea70f27f-0c50-41b6-a711-435eb087193a" (UID: "ea70f27f-0c50-41b6-a711-435eb087193a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.793052 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5" (OuterVolumeSpecName: "kube-api-access-7rtb5") pod "ea70f27f-0c50-41b6-a711-435eb087193a" (UID: "ea70f27f-0c50-41b6-a711-435eb087193a"). InnerVolumeSpecName "kube-api-access-7rtb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.831467 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea70f27f-0c50-41b6-a711-435eb087193a" containerID="81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc" exitCode=0 Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.831518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerDied","Data":"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc"} Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.831547 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hh5s" event={"ID":"ea70f27f-0c50-41b6-a711-435eb087193a","Type":"ContainerDied","Data":"a758826dfd5f624de2c8c2d385bae4e5588c5fefb61b12da375acab68645acfd"} Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.831564 4754 scope.go:117] "RemoveContainer" containerID="81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.831731 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hh5s" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.866001 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea70f27f-0c50-41b6-a711-435eb087193a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.866029 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rtb5\" (UniqueName: \"kubernetes.io/projected/ea70f27f-0c50-41b6-a711-435eb087193a-kube-api-access-7rtb5\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.875695 4754 scope.go:117] "RemoveContainer" containerID="0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.891282 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.901522 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hh5s"] Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.904536 4754 scope.go:117] "RemoveContainer" containerID="55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.948497 4754 scope.go:117] "RemoveContainer" containerID="81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc" Oct 11 04:08:53 crc kubenswrapper[4754]: E1011 04:08:53.949263 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc\": container with ID starting with 81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc not found: ID does not exist" containerID="81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.949306 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc"} err="failed to get container status \"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc\": rpc error: code = NotFound desc = could not find container \"81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc\": container with ID starting with 81b881f688efa5ad5d4808957989e759eee2fed73368b8b2125652eec24d78fc not found: ID does not exist" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.949329 4754 scope.go:117] "RemoveContainer" containerID="0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32" Oct 11 04:08:53 crc kubenswrapper[4754]: E1011 04:08:53.949809 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32\": container with ID starting with 0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32 not found: ID does not exist" containerID="0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.949828 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32"} err="failed to get container status \"0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32\": rpc error: code = NotFound desc = could not find container \"0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32\": container with ID starting with 0e28a2dd72c7c90b422573ed7d4bda389eb888c18de5a2ceaf1d9d478f7ebf32 not found: ID does not exist" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.949840 4754 scope.go:117] "RemoveContainer" containerID="55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c" Oct 11 04:08:53 crc kubenswrapper[4754]: E1011 04:08:53.950311 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c\": container with ID starting with 55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c not found: ID does not exist" containerID="55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c" Oct 11 04:08:53 crc kubenswrapper[4754]: I1011 04:08:53.950388 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c"} err="failed to get container status \"55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c\": rpc error: code = NotFound desc = could not find container \"55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c\": container with ID starting with 55ed1b8fe817943b5378f4eb8fe6bcdd3cba116101089e6c84b9caf09fb1501c not found: ID does not exist" Oct 11 04:08:55 crc kubenswrapper[4754]: I1011 04:08:55.094454 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" path="/var/lib/kubelet/pods/ea70f27f-0c50-41b6-a711-435eb087193a/volumes" Oct 11 04:09:00 crc kubenswrapper[4754]: I1011 04:09:00.736894 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:09:00 crc kubenswrapper[4754]: I1011 04:09:00.737489 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:09:30 crc kubenswrapper[4754]: I1011 04:09:30.737098 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:09:30 crc kubenswrapper[4754]: I1011 04:09:30.739806 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:09:30 crc kubenswrapper[4754]: I1011 04:09:30.740002 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:09:30 crc kubenswrapper[4754]: I1011 04:09:30.741338 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:09:30 crc kubenswrapper[4754]: I1011 04:09:30.741538 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" gracePeriod=600 Oct 11 04:09:30 crc kubenswrapper[4754]: E1011 04:09:30.880322 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:09:31 crc kubenswrapper[4754]: I1011 04:09:31.250881 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" exitCode=0 Oct 11 04:09:31 crc kubenswrapper[4754]: I1011 04:09:31.250900 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710"} Oct 11 04:09:31 crc kubenswrapper[4754]: I1011 04:09:31.251532 4754 scope.go:117] "RemoveContainer" containerID="fc8d8d30670fd80e7ef414d669098e3222744716045a21e152fb83d230c17ff6" Oct 11 04:09:31 crc kubenswrapper[4754]: I1011 04:09:31.252434 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:09:31 crc kubenswrapper[4754]: E1011 04:09:31.252730 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:09:42 crc kubenswrapper[4754]: I1011 04:09:42.083860 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:09:42 crc kubenswrapper[4754]: E1011 04:09:42.085039 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:09:55 crc kubenswrapper[4754]: I1011 04:09:55.088303 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:09:55 crc kubenswrapper[4754]: E1011 04:09:55.089256 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:08 crc kubenswrapper[4754]: I1011 04:10:08.083383 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:10:08 crc kubenswrapper[4754]: E1011 04:10:08.084457 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:22 crc kubenswrapper[4754]: I1011 04:10:22.083579 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:10:22 crc kubenswrapper[4754]: E1011 04:10:22.084588 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:33 crc kubenswrapper[4754]: I1011 04:10:33.091389 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:10:33 crc kubenswrapper[4754]: E1011 04:10:33.092304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.462041 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.463938 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464134 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.464242 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="extract-utilities" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464310 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="extract-utilities" Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.464367 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464420 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.464501 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="extract-utilities" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464585 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="extract-utilities" Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.464660 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="extract-content" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464718 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="extract-content" Oct 11 04:10:46 crc kubenswrapper[4754]: E1011 04:10:46.464781 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="extract-content" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.464833 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="extract-content" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.465150 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb8cab65-7c24-4c8b-9e60-4134f080c17a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.465268 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea70f27f-0c50-41b6-a711-435eb087193a" containerName="registry-server" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.466911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.484899 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.641932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.642034 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.642186 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46jz8\" (UniqueName: \"kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.648001 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.656480 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.670922 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.744067 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.744126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.744265 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46jz8\" (UniqueName: \"kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.744576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.744763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.769325 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46jz8\" (UniqueName: \"kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8\") pod \"redhat-operators-t8dmj\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.791562 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.847799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfxm5\" (UniqueName: \"kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.848311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.848371 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.950512 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfxm5\" (UniqueName: \"kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.950606 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.950647 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.951459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.951684 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:46 crc kubenswrapper[4754]: I1011 04:10:46.993643 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfxm5\" (UniqueName: \"kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5\") pod \"community-operators-msvkg\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:47 crc kubenswrapper[4754]: I1011 04:10:47.086383 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:10:47 crc kubenswrapper[4754]: E1011 04:10:47.087090 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:47 crc kubenswrapper[4754]: I1011 04:10:47.273512 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:10:47 crc kubenswrapper[4754]: I1011 04:10:47.380103 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:10:47 crc kubenswrapper[4754]: W1011 04:10:47.885122 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod076943bd_d7b8_4569_8e76_ad5495150eaa.slice/crio-c1d28d2640e036deb12bc26acdfd3b19846aa486527740155a1d4b7b57fa89ae WatchSource:0}: Error finding container c1d28d2640e036deb12bc26acdfd3b19846aa486527740155a1d4b7b57fa89ae: Status 404 returned error can't find the container with id c1d28d2640e036deb12bc26acdfd3b19846aa486527740155a1d4b7b57fa89ae Oct 11 04:10:47 crc kubenswrapper[4754]: I1011 04:10:47.895893 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:10:48 crc kubenswrapper[4754]: I1011 04:10:48.018886 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerStarted","Data":"c1d28d2640e036deb12bc26acdfd3b19846aa486527740155a1d4b7b57fa89ae"} Oct 11 04:10:48 crc kubenswrapper[4754]: I1011 04:10:48.021316 4754 generic.go:334] "Generic (PLEG): container finished" podID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerID="5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c" exitCode=0 Oct 11 04:10:48 crc kubenswrapper[4754]: I1011 04:10:48.021455 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerDied","Data":"5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c"} Oct 11 04:10:48 crc kubenswrapper[4754]: I1011 04:10:48.022469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerStarted","Data":"5d84532596f42493a5dfc62d4880b4d76f023197c7e1b63fdf10c31720b4b4ab"} Oct 11 04:10:49 crc kubenswrapper[4754]: I1011 04:10:49.037004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerStarted","Data":"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422"} Oct 11 04:10:49 crc kubenswrapper[4754]: I1011 04:10:49.041386 4754 generic.go:334] "Generic (PLEG): container finished" podID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerID="43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d" exitCode=0 Oct 11 04:10:49 crc kubenswrapper[4754]: I1011 04:10:49.041434 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerDied","Data":"43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d"} Oct 11 04:10:50 crc kubenswrapper[4754]: I1011 04:10:50.054361 4754 generic.go:334] "Generic (PLEG): container finished" podID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerID="9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422" exitCode=0 Oct 11 04:10:50 crc kubenswrapper[4754]: I1011 04:10:50.054445 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerDied","Data":"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422"} Oct 11 04:10:51 crc kubenswrapper[4754]: I1011 04:10:51.065685 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerStarted","Data":"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187"} Oct 11 04:10:51 crc kubenswrapper[4754]: I1011 04:10:51.098168 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t8dmj" podStartSLOduration=2.618934585 podStartE2EDuration="5.098137634s" podCreationTimestamp="2025-10-11 04:10:46 +0000 UTC" firstStartedPulling="2025-10-11 04:10:48.023411116 +0000 UTC m=+3895.582355911" lastFinishedPulling="2025-10-11 04:10:50.502614175 +0000 UTC m=+3898.061558960" observedRunningTime="2025-10-11 04:10:51.094991086 +0000 UTC m=+3898.653935891" watchObservedRunningTime="2025-10-11 04:10:51.098137634 +0000 UTC m=+3898.657082429" Oct 11 04:10:56 crc kubenswrapper[4754]: I1011 04:10:56.792789 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:56 crc kubenswrapper[4754]: I1011 04:10:56.793864 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:56 crc kubenswrapper[4754]: I1011 04:10:56.849593 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:57 crc kubenswrapper[4754]: I1011 04:10:57.210390 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:57 crc kubenswrapper[4754]: I1011 04:10:57.351709 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:10:58 crc kubenswrapper[4754]: I1011 04:10:58.084626 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:10:58 crc kubenswrapper[4754]: E1011 04:10:58.085576 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.177840 4754 generic.go:334] "Generic (PLEG): container finished" podID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerID="07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d" exitCode=0 Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.177926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerDied","Data":"07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d"} Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.180684 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t8dmj" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="registry-server" containerID="cri-o://a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187" gracePeriod=2 Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.846561 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.884797 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities\") pod \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.884848 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46jz8\" (UniqueName: \"kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8\") pod \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.884865 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content\") pod \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\" (UID: \"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3\") " Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.885783 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities" (OuterVolumeSpecName: "utilities") pod "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" (UID: "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.896794 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8" (OuterVolumeSpecName: "kube-api-access-46jz8") pod "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" (UID: "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3"). InnerVolumeSpecName "kube-api-access-46jz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.987279 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:59 crc kubenswrapper[4754]: I1011 04:10:59.987618 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46jz8\" (UniqueName: \"kubernetes.io/projected/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-kube-api-access-46jz8\") on node \"crc\" DevicePath \"\"" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.204492 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerStarted","Data":"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2"} Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.207525 4754 generic.go:334] "Generic (PLEG): container finished" podID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerID="a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187" exitCode=0 Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.207588 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8dmj" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.207587 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerDied","Data":"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187"} Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.207935 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8dmj" event={"ID":"a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3","Type":"ContainerDied","Data":"5d84532596f42493a5dfc62d4880b4d76f023197c7e1b63fdf10c31720b4b4ab"} Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.207985 4754 scope.go:117] "RemoveContainer" containerID="a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.230026 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-msvkg" podStartSLOduration=3.573245263 podStartE2EDuration="14.230006517s" podCreationTimestamp="2025-10-11 04:10:46 +0000 UTC" firstStartedPulling="2025-10-11 04:10:49.044236101 +0000 UTC m=+3896.603180886" lastFinishedPulling="2025-10-11 04:10:59.700997355 +0000 UTC m=+3907.259942140" observedRunningTime="2025-10-11 04:11:00.226488189 +0000 UTC m=+3907.785432984" watchObservedRunningTime="2025-10-11 04:11:00.230006517 +0000 UTC m=+3907.788951302" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.239482 4754 scope.go:117] "RemoveContainer" containerID="9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.281348 4754 scope.go:117] "RemoveContainer" containerID="5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.326045 4754 scope.go:117] "RemoveContainer" containerID="a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187" Oct 11 04:11:00 crc kubenswrapper[4754]: E1011 04:11:00.326668 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187\": container with ID starting with a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187 not found: ID does not exist" containerID="a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.326732 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187"} err="failed to get container status \"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187\": rpc error: code = NotFound desc = could not find container \"a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187\": container with ID starting with a190d2f30403a6c6a4d729ab9dcf86d31d60ff37d7937764fb6193af7b459187 not found: ID does not exist" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.326769 4754 scope.go:117] "RemoveContainer" containerID="9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422" Oct 11 04:11:00 crc kubenswrapper[4754]: E1011 04:11:00.327073 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422\": container with ID starting with 9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422 not found: ID does not exist" containerID="9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.327108 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422"} err="failed to get container status \"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422\": rpc error: code = NotFound desc = could not find container \"9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422\": container with ID starting with 9ccdf55bce1c3c064284b0a1adedf84c632ae26dbad9ced5f2551a00a94f0422 not found: ID does not exist" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.327129 4754 scope.go:117] "RemoveContainer" containerID="5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c" Oct 11 04:11:00 crc kubenswrapper[4754]: E1011 04:11:00.327367 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c\": container with ID starting with 5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c not found: ID does not exist" containerID="5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.327402 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c"} err="failed to get container status \"5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c\": rpc error: code = NotFound desc = could not find container \"5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c\": container with ID starting with 5a4cab008d9bdf34bc294a4218109c0fcf3b4b1f5e8d7301c5cef9aa6fd7fe6c not found: ID does not exist" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.496483 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" (UID: "a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.526476 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.542989 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:11:00 crc kubenswrapper[4754]: I1011 04:11:00.554710 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t8dmj"] Oct 11 04:11:01 crc kubenswrapper[4754]: I1011 04:11:01.096312 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" path="/var/lib/kubelet/pods/a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3/volumes" Oct 11 04:11:07 crc kubenswrapper[4754]: I1011 04:11:07.275366 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:07 crc kubenswrapper[4754]: I1011 04:11:07.277559 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:07 crc kubenswrapper[4754]: I1011 04:11:07.396498 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:08 crc kubenswrapper[4754]: I1011 04:11:08.371387 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:08 crc kubenswrapper[4754]: I1011 04:11:08.469377 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:11:09 crc kubenswrapper[4754]: I1011 04:11:09.083762 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:11:09 crc kubenswrapper[4754]: E1011 04:11:09.084442 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:11:10 crc kubenswrapper[4754]: I1011 04:11:10.328327 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-msvkg" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="registry-server" containerID="cri-o://f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2" gracePeriod=2 Oct 11 04:11:10 crc kubenswrapper[4754]: I1011 04:11:10.961585 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.016478 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfxm5\" (UniqueName: \"kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5\") pod \"076943bd-d7b8-4569-8e76-ad5495150eaa\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.016726 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities\") pod \"076943bd-d7b8-4569-8e76-ad5495150eaa\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.016764 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content\") pod \"076943bd-d7b8-4569-8e76-ad5495150eaa\" (UID: \"076943bd-d7b8-4569-8e76-ad5495150eaa\") " Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.019896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities" (OuterVolumeSpecName: "utilities") pod "076943bd-d7b8-4569-8e76-ad5495150eaa" (UID: "076943bd-d7b8-4569-8e76-ad5495150eaa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.025536 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5" (OuterVolumeSpecName: "kube-api-access-gfxm5") pod "076943bd-d7b8-4569-8e76-ad5495150eaa" (UID: "076943bd-d7b8-4569-8e76-ad5495150eaa"). InnerVolumeSpecName "kube-api-access-gfxm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.067444 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "076943bd-d7b8-4569-8e76-ad5495150eaa" (UID: "076943bd-d7b8-4569-8e76-ad5495150eaa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.119892 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfxm5\" (UniqueName: \"kubernetes.io/projected/076943bd-d7b8-4569-8e76-ad5495150eaa-kube-api-access-gfxm5\") on node \"crc\" DevicePath \"\"" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.119931 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.119941 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076943bd-d7b8-4569-8e76-ad5495150eaa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.339364 4754 generic.go:334] "Generic (PLEG): container finished" podID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerID="f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2" exitCode=0 Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.339409 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerDied","Data":"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2"} Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.339446 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msvkg" event={"ID":"076943bd-d7b8-4569-8e76-ad5495150eaa","Type":"ContainerDied","Data":"c1d28d2640e036deb12bc26acdfd3b19846aa486527740155a1d4b7b57fa89ae"} Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.339466 4754 scope.go:117] "RemoveContainer" containerID="f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.339474 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msvkg" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.369587 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.376771 4754 scope.go:117] "RemoveContainer" containerID="07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.382076 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-msvkg"] Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.401950 4754 scope.go:117] "RemoveContainer" containerID="43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.455006 4754 scope.go:117] "RemoveContainer" containerID="f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2" Oct 11 04:11:11 crc kubenswrapper[4754]: E1011 04:11:11.455571 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2\": container with ID starting with f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2 not found: ID does not exist" containerID="f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.455621 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2"} err="failed to get container status \"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2\": rpc error: code = NotFound desc = could not find container \"f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2\": container with ID starting with f16ff466b62b3d90ddb1a0c4bcb7b7003ff77ad348150e28fad29cf4fa35cdf2 not found: ID does not exist" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.455659 4754 scope.go:117] "RemoveContainer" containerID="07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d" Oct 11 04:11:11 crc kubenswrapper[4754]: E1011 04:11:11.456093 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d\": container with ID starting with 07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d not found: ID does not exist" containerID="07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.456136 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d"} err="failed to get container status \"07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d\": rpc error: code = NotFound desc = could not find container \"07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d\": container with ID starting with 07440bd170e20f56cb42f256b18f799dbaa804ebf445a3b26d670175b942e23d not found: ID does not exist" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.456167 4754 scope.go:117] "RemoveContainer" containerID="43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d" Oct 11 04:11:11 crc kubenswrapper[4754]: E1011 04:11:11.456568 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d\": container with ID starting with 43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d not found: ID does not exist" containerID="43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d" Oct 11 04:11:11 crc kubenswrapper[4754]: I1011 04:11:11.456594 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d"} err="failed to get container status \"43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d\": rpc error: code = NotFound desc = could not find container \"43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d\": container with ID starting with 43225551537f5581f22eade2ce3ff46d9319ce14853e4775b3968bed020fe55d not found: ID does not exist" Oct 11 04:11:13 crc kubenswrapper[4754]: I1011 04:11:13.108979 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" path="/var/lib/kubelet/pods/076943bd-d7b8-4569-8e76-ad5495150eaa/volumes" Oct 11 04:11:20 crc kubenswrapper[4754]: I1011 04:11:20.084100 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:11:20 crc kubenswrapper[4754]: E1011 04:11:20.085275 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:11:31 crc kubenswrapper[4754]: I1011 04:11:31.088754 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:11:31 crc kubenswrapper[4754]: E1011 04:11:31.089586 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:11:45 crc kubenswrapper[4754]: I1011 04:11:45.084685 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:11:45 crc kubenswrapper[4754]: E1011 04:11:45.088036 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:11:56 crc kubenswrapper[4754]: I1011 04:11:56.083785 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:11:56 crc kubenswrapper[4754]: E1011 04:11:56.085066 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:12:10 crc kubenswrapper[4754]: I1011 04:12:10.083916 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:12:10 crc kubenswrapper[4754]: E1011 04:12:10.084902 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:12:24 crc kubenswrapper[4754]: I1011 04:12:24.084351 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:12:24 crc kubenswrapper[4754]: E1011 04:12:24.085207 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:12:35 crc kubenswrapper[4754]: I1011 04:12:35.087469 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:12:35 crc kubenswrapper[4754]: E1011 04:12:35.089213 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:12:46 crc kubenswrapper[4754]: I1011 04:12:46.083505 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:12:46 crc kubenswrapper[4754]: E1011 04:12:46.084433 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:12:57 crc kubenswrapper[4754]: I1011 04:12:57.084326 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:12:57 crc kubenswrapper[4754]: E1011 04:12:57.085211 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:13:09 crc kubenswrapper[4754]: I1011 04:13:09.084501 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:13:09 crc kubenswrapper[4754]: E1011 04:13:09.086849 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:13:20 crc kubenswrapper[4754]: I1011 04:13:20.083809 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:13:20 crc kubenswrapper[4754]: E1011 04:13:20.084993 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:13:33 crc kubenswrapper[4754]: I1011 04:13:33.091146 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:13:33 crc kubenswrapper[4754]: E1011 04:13:33.092300 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:13:47 crc kubenswrapper[4754]: I1011 04:13:47.085148 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:13:47 crc kubenswrapper[4754]: E1011 04:13:47.086007 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:14:02 crc kubenswrapper[4754]: I1011 04:14:02.083469 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:14:02 crc kubenswrapper[4754]: E1011 04:14:02.084798 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:14:15 crc kubenswrapper[4754]: I1011 04:14:15.083634 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:14:15 crc kubenswrapper[4754]: E1011 04:14:15.084587 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:14:29 crc kubenswrapper[4754]: I1011 04:14:29.084733 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:14:29 crc kubenswrapper[4754]: E1011 04:14:29.085719 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:14:40 crc kubenswrapper[4754]: I1011 04:14:40.083394 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:14:40 crc kubenswrapper[4754]: I1011 04:14:40.554918 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794"} Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.154082 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk"] Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155282 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="extract-content" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155301 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="extract-content" Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155329 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="extract-utilities" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155336 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="extract-utilities" Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155348 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="extract-utilities" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155354 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="extract-utilities" Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155365 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155371 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155419 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155425 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: E1011 04:15:00.155440 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="extract-content" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155447 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="extract-content" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155675 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a07c9cc4-a92b-48b4-bd2a-03b5080f8bd3" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.155695 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="076943bd-d7b8-4569-8e76-ad5495150eaa" containerName="registry-server" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.156507 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.160479 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.163458 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.178056 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk"] Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.290361 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdxfr\" (UniqueName: \"kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.290409 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.290484 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.393505 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdxfr\" (UniqueName: \"kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.394120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.394431 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.395808 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.404673 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.416162 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdxfr\" (UniqueName: \"kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr\") pod \"collect-profiles-29335935-zs4gk\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.516087 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:00 crc kubenswrapper[4754]: I1011 04:15:00.982115 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk"] Oct 11 04:15:01 crc kubenswrapper[4754]: W1011 04:15:01.384655 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2743361e_eb5c_416c_b925_1aa7d057750d.slice/crio-0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90 WatchSource:0}: Error finding container 0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90: Status 404 returned error can't find the container with id 0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90 Oct 11 04:15:01 crc kubenswrapper[4754]: I1011 04:15:01.814040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" event={"ID":"2743361e-eb5c-416c-b925-1aa7d057750d","Type":"ContainerStarted","Data":"533eee567cb45767006bd64fd2725a629c500b810f7be6764961345c7cd65665"} Oct 11 04:15:01 crc kubenswrapper[4754]: I1011 04:15:01.814694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" event={"ID":"2743361e-eb5c-416c-b925-1aa7d057750d","Type":"ContainerStarted","Data":"0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90"} Oct 11 04:15:01 crc kubenswrapper[4754]: I1011 04:15:01.836743 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" podStartSLOduration=1.836714679 podStartE2EDuration="1.836714679s" podCreationTimestamp="2025-10-11 04:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:15:01.834640001 +0000 UTC m=+4149.393584786" watchObservedRunningTime="2025-10-11 04:15:01.836714679 +0000 UTC m=+4149.395659474" Oct 11 04:15:02 crc kubenswrapper[4754]: I1011 04:15:02.826782 4754 generic.go:334] "Generic (PLEG): container finished" podID="2743361e-eb5c-416c-b925-1aa7d057750d" containerID="533eee567cb45767006bd64fd2725a629c500b810f7be6764961345c7cd65665" exitCode=0 Oct 11 04:15:02 crc kubenswrapper[4754]: I1011 04:15:02.826848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" event={"ID":"2743361e-eb5c-416c-b925-1aa7d057750d","Type":"ContainerDied","Data":"533eee567cb45767006bd64fd2725a629c500b810f7be6764961345c7cd65665"} Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.499586 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.651303 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume\") pod \"2743361e-eb5c-416c-b925-1aa7d057750d\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.651471 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume\") pod \"2743361e-eb5c-416c-b925-1aa7d057750d\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.651629 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdxfr\" (UniqueName: \"kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr\") pod \"2743361e-eb5c-416c-b925-1aa7d057750d\" (UID: \"2743361e-eb5c-416c-b925-1aa7d057750d\") " Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.652910 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume" (OuterVolumeSpecName: "config-volume") pod "2743361e-eb5c-416c-b925-1aa7d057750d" (UID: "2743361e-eb5c-416c-b925-1aa7d057750d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.662363 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2743361e-eb5c-416c-b925-1aa7d057750d" (UID: "2743361e-eb5c-416c-b925-1aa7d057750d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.663231 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr" (OuterVolumeSpecName: "kube-api-access-qdxfr") pod "2743361e-eb5c-416c-b925-1aa7d057750d" (UID: "2743361e-eb5c-416c-b925-1aa7d057750d"). InnerVolumeSpecName "kube-api-access-qdxfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.754544 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2743361e-eb5c-416c-b925-1aa7d057750d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.754596 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2743361e-eb5c-416c-b925-1aa7d057750d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.754615 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdxfr\" (UniqueName: \"kubernetes.io/projected/2743361e-eb5c-416c-b925-1aa7d057750d-kube-api-access-qdxfr\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.852738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" event={"ID":"2743361e-eb5c-416c-b925-1aa7d057750d","Type":"ContainerDied","Data":"0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90"} Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.852817 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0adebb3bfb192e77ebc4f3f33222cb223431d7e8c0ca83b9e89dcca3f0534c90" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.852944 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-zs4gk" Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.930611 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv"] Oct 11 04:15:04 crc kubenswrapper[4754]: I1011 04:15:04.941807 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335890-q8csv"] Oct 11 04:15:05 crc kubenswrapper[4754]: I1011 04:15:05.123815 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d842ba1d-77ef-4397-a138-849800aff4a9" path="/var/lib/kubelet/pods/d842ba1d-77ef-4397-a138-849800aff4a9/volumes" Oct 11 04:15:07 crc kubenswrapper[4754]: I1011 04:15:07.671173 4754 scope.go:117] "RemoveContainer" containerID="862993f30a12e79f8b52b2ac44dabc65baa9d1a7082832c6d3feb510cd1d0945" Oct 11 04:17:00 crc kubenswrapper[4754]: I1011 04:17:00.737417 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:17:00 crc kubenswrapper[4754]: I1011 04:17:00.738526 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:17:30 crc kubenswrapper[4754]: I1011 04:17:30.736650 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:17:30 crc kubenswrapper[4754]: I1011 04:17:30.737774 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:18:00 crc kubenswrapper[4754]: I1011 04:18:00.736722 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:18:00 crc kubenswrapper[4754]: I1011 04:18:00.737775 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:18:00 crc kubenswrapper[4754]: I1011 04:18:00.737870 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:18:00 crc kubenswrapper[4754]: I1011 04:18:00.739370 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:18:00 crc kubenswrapper[4754]: I1011 04:18:00.739529 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794" gracePeriod=600 Oct 11 04:18:01 crc kubenswrapper[4754]: I1011 04:18:01.762511 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794" exitCode=0 Oct 11 04:18:01 crc kubenswrapper[4754]: I1011 04:18:01.762553 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794"} Oct 11 04:18:01 crc kubenswrapper[4754]: I1011 04:18:01.764076 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc"} Oct 11 04:18:01 crc kubenswrapper[4754]: I1011 04:18:01.764182 4754 scope.go:117] "RemoveContainer" containerID="51effd34d99d73a4a7b9e61dd476389ccdbde81ae3369402810f18cd24c3c710" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.569418 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:21 crc kubenswrapper[4754]: E1011 04:19:21.570462 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2743361e-eb5c-416c-b925-1aa7d057750d" containerName="collect-profiles" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.570477 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2743361e-eb5c-416c-b925-1aa7d057750d" containerName="collect-profiles" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.570700 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2743361e-eb5c-416c-b925-1aa7d057750d" containerName="collect-profiles" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.572503 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.582028 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.758950 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmjg\" (UniqueName: \"kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.759101 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.759893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.862169 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.862273 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmjg\" (UniqueName: \"kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.862309 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.862828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:21 crc kubenswrapper[4754]: I1011 04:19:21.862917 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:22 crc kubenswrapper[4754]: I1011 04:19:22.072741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmjg\" (UniqueName: \"kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg\") pod \"redhat-marketplace-76fq6\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:22 crc kubenswrapper[4754]: I1011 04:19:22.198672 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:22 crc kubenswrapper[4754]: I1011 04:19:22.749675 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:23 crc kubenswrapper[4754]: I1011 04:19:23.602389 4754 generic.go:334] "Generic (PLEG): container finished" podID="06ae0104-8bef-48e9-b018-a072786a32f2" containerID="7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923" exitCode=0 Oct 11 04:19:23 crc kubenswrapper[4754]: I1011 04:19:23.602515 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerDied","Data":"7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923"} Oct 11 04:19:23 crc kubenswrapper[4754]: I1011 04:19:23.602744 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerStarted","Data":"95355e6af8570e9cdbbc8aa53f5b8d3554543ccbae726b19c76e9e86270324bd"} Oct 11 04:19:23 crc kubenswrapper[4754]: I1011 04:19:23.605123 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:19:24 crc kubenswrapper[4754]: I1011 04:19:24.614134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerStarted","Data":"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0"} Oct 11 04:19:25 crc kubenswrapper[4754]: I1011 04:19:25.633229 4754 generic.go:334] "Generic (PLEG): container finished" podID="06ae0104-8bef-48e9-b018-a072786a32f2" containerID="ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0" exitCode=0 Oct 11 04:19:25 crc kubenswrapper[4754]: I1011 04:19:25.633339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerDied","Data":"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0"} Oct 11 04:19:26 crc kubenswrapper[4754]: I1011 04:19:26.644895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerStarted","Data":"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f"} Oct 11 04:19:26 crc kubenswrapper[4754]: I1011 04:19:26.668816 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-76fq6" podStartSLOduration=3.08332522 podStartE2EDuration="5.668794093s" podCreationTimestamp="2025-10-11 04:19:21 +0000 UTC" firstStartedPulling="2025-10-11 04:19:23.604641506 +0000 UTC m=+4411.163586341" lastFinishedPulling="2025-10-11 04:19:26.190110429 +0000 UTC m=+4413.749055214" observedRunningTime="2025-10-11 04:19:26.661271701 +0000 UTC m=+4414.220216496" watchObservedRunningTime="2025-10-11 04:19:26.668794093 +0000 UTC m=+4414.227738878" Oct 11 04:19:32 crc kubenswrapper[4754]: I1011 04:19:32.199192 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:32 crc kubenswrapper[4754]: I1011 04:19:32.200114 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:32 crc kubenswrapper[4754]: I1011 04:19:32.427159 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:32 crc kubenswrapper[4754]: I1011 04:19:32.762471 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:32 crc kubenswrapper[4754]: I1011 04:19:32.810845 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:34 crc kubenswrapper[4754]: I1011 04:19:34.733318 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-76fq6" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="registry-server" containerID="cri-o://3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f" gracePeriod=2 Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.203304 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.349510 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcmjg\" (UniqueName: \"kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg\") pod \"06ae0104-8bef-48e9-b018-a072786a32f2\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.349701 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities\") pod \"06ae0104-8bef-48e9-b018-a072786a32f2\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.349752 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content\") pod \"06ae0104-8bef-48e9-b018-a072786a32f2\" (UID: \"06ae0104-8bef-48e9-b018-a072786a32f2\") " Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.354546 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities" (OuterVolumeSpecName: "utilities") pod "06ae0104-8bef-48e9-b018-a072786a32f2" (UID: "06ae0104-8bef-48e9-b018-a072786a32f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.361631 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg" (OuterVolumeSpecName: "kube-api-access-wcmjg") pod "06ae0104-8bef-48e9-b018-a072786a32f2" (UID: "06ae0104-8bef-48e9-b018-a072786a32f2"). InnerVolumeSpecName "kube-api-access-wcmjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.365021 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06ae0104-8bef-48e9-b018-a072786a32f2" (UID: "06ae0104-8bef-48e9-b018-a072786a32f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.452441 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcmjg\" (UniqueName: \"kubernetes.io/projected/06ae0104-8bef-48e9-b018-a072786a32f2-kube-api-access-wcmjg\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.452763 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.452777 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ae0104-8bef-48e9-b018-a072786a32f2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.745603 4754 generic.go:334] "Generic (PLEG): container finished" podID="06ae0104-8bef-48e9-b018-a072786a32f2" containerID="3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f" exitCode=0 Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.745668 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerDied","Data":"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f"} Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.745714 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76fq6" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.745732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76fq6" event={"ID":"06ae0104-8bef-48e9-b018-a072786a32f2","Type":"ContainerDied","Data":"95355e6af8570e9cdbbc8aa53f5b8d3554543ccbae726b19c76e9e86270324bd"} Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.745753 4754 scope.go:117] "RemoveContainer" containerID="3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.773315 4754 scope.go:117] "RemoveContainer" containerID="ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.786198 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.798957 4754 scope.go:117] "RemoveContainer" containerID="7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.799611 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-76fq6"] Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.902575 4754 scope.go:117] "RemoveContainer" containerID="3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f" Oct 11 04:19:35 crc kubenswrapper[4754]: E1011 04:19:35.903045 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f\": container with ID starting with 3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f not found: ID does not exist" containerID="3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.903077 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f"} err="failed to get container status \"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f\": rpc error: code = NotFound desc = could not find container \"3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f\": container with ID starting with 3c2aa69cebc01e4576b3bd6680a7c41c0b1ef39b8e8689766200bb7d8e34247f not found: ID does not exist" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.903098 4754 scope.go:117] "RemoveContainer" containerID="ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0" Oct 11 04:19:35 crc kubenswrapper[4754]: E1011 04:19:35.903632 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0\": container with ID starting with ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0 not found: ID does not exist" containerID="ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.903696 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0"} err="failed to get container status \"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0\": rpc error: code = NotFound desc = could not find container \"ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0\": container with ID starting with ec44bea0c1a6de5dfe5f5e418d1d0683830ef399d531c83f55d98e38eedc88b0 not found: ID does not exist" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.903737 4754 scope.go:117] "RemoveContainer" containerID="7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923" Oct 11 04:19:35 crc kubenswrapper[4754]: E1011 04:19:35.904820 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923\": container with ID starting with 7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923 not found: ID does not exist" containerID="7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923" Oct 11 04:19:35 crc kubenswrapper[4754]: I1011 04:19:35.904846 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923"} err="failed to get container status \"7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923\": rpc error: code = NotFound desc = could not find container \"7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923\": container with ID starting with 7815e5e1af4bc54700909fd92688c18aac53c99726f7cf50b1c1ca439683b923 not found: ID does not exist" Oct 11 04:19:37 crc kubenswrapper[4754]: I1011 04:19:37.096938 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" path="/var/lib/kubelet/pods/06ae0104-8bef-48e9-b018-a072786a32f2/volumes" Oct 11 04:20:30 crc kubenswrapper[4754]: I1011 04:20:30.736160 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:20:30 crc kubenswrapper[4754]: I1011 04:20:30.737018 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:21:00 crc kubenswrapper[4754]: I1011 04:21:00.736117 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:21:00 crc kubenswrapper[4754]: I1011 04:21:00.736716 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.737141 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.737925 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.738063 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.739278 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.739370 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" gracePeriod=600 Oct 11 04:21:30 crc kubenswrapper[4754]: E1011 04:21:30.875828 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.939054 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" exitCode=0 Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.939120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc"} Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.939174 4754 scope.go:117] "RemoveContainer" containerID="93760c264cd872f38ade7a7b39a5e007cb6db8c3b495d365c2b717583fd0b794" Oct 11 04:21:30 crc kubenswrapper[4754]: I1011 04:21:30.939949 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:21:30 crc kubenswrapper[4754]: E1011 04:21:30.940348 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:21:46 crc kubenswrapper[4754]: I1011 04:21:46.083722 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:21:46 crc kubenswrapper[4754]: E1011 04:21:46.084787 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:22:01 crc kubenswrapper[4754]: I1011 04:22:01.084384 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:22:01 crc kubenswrapper[4754]: E1011 04:22:01.086589 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:22:14 crc kubenswrapper[4754]: I1011 04:22:14.083980 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:22:14 crc kubenswrapper[4754]: E1011 04:22:14.085242 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:22:28 crc kubenswrapper[4754]: I1011 04:22:28.083231 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:22:28 crc kubenswrapper[4754]: E1011 04:22:28.083988 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:22:40 crc kubenswrapper[4754]: I1011 04:22:40.086383 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:22:40 crc kubenswrapper[4754]: E1011 04:22:40.088479 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:22:52 crc kubenswrapper[4754]: I1011 04:22:52.084122 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:22:52 crc kubenswrapper[4754]: E1011 04:22:52.085119 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:23:06 crc kubenswrapper[4754]: I1011 04:23:06.083623 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:23:06 crc kubenswrapper[4754]: E1011 04:23:06.084323 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.637620 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:13 crc kubenswrapper[4754]: E1011 04:23:13.638947 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="extract-content" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.638991 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="extract-content" Oct 11 04:23:13 crc kubenswrapper[4754]: E1011 04:23:13.639015 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="registry-server" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.639026 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="registry-server" Oct 11 04:23:13 crc kubenswrapper[4754]: E1011 04:23:13.639046 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="extract-utilities" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.639055 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="extract-utilities" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.639353 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ae0104-8bef-48e9-b018-a072786a32f2" containerName="registry-server" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.641206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.657937 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.778166 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncbgx\" (UniqueName: \"kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.778232 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.778329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.881686 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncbgx\" (UniqueName: \"kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.881754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.881876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.882381 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.882409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.908391 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncbgx\" (UniqueName: \"kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx\") pod \"certified-operators-8w7nm\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:13 crc kubenswrapper[4754]: I1011 04:23:13.968758 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:14 crc kubenswrapper[4754]: I1011 04:23:14.590847 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:14 crc kubenswrapper[4754]: I1011 04:23:14.925469 4754 generic.go:334] "Generic (PLEG): container finished" podID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerID="446efa27d7357c28d432336855f4e1ed04c4c19a6071e6869913db5183a3c940" exitCode=0 Oct 11 04:23:14 crc kubenswrapper[4754]: I1011 04:23:14.925705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerDied","Data":"446efa27d7357c28d432336855f4e1ed04c4c19a6071e6869913db5183a3c940"} Oct 11 04:23:14 crc kubenswrapper[4754]: I1011 04:23:14.925914 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerStarted","Data":"73bfaccfa39be9d811163eb095eb84d9749e39abd79cc4678c32b7f3c31681cb"} Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.424871 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.427560 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.446892 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.522124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xggds\" (UniqueName: \"kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.522184 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.522266 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.623908 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xggds\" (UniqueName: \"kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.623990 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.624085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.624718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.624878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.652034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xggds\" (UniqueName: \"kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds\") pod \"redhat-operators-vk9r2\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:15 crc kubenswrapper[4754]: I1011 04:23:15.810242 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.026217 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.028711 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.038445 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.139204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb2cs\" (UniqueName: \"kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.139379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.139449 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.241876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb2cs\" (UniqueName: \"kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.242200 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.242321 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.242872 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.243210 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.270446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb2cs\" (UniqueName: \"kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs\") pod \"community-operators-p6sh7\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.645766 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.809364 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.949129 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerStarted","Data":"abcb3d1738cc69ee2b12ddb6f2ccd2ea02380c139a0bdaf3fda8b6ac01ea0dc7"} Oct 11 04:23:16 crc kubenswrapper[4754]: I1011 04:23:16.959895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerStarted","Data":"527bc5e4bb7229fcce4dafe1f89b56d1663c5ed290d23cf8a90439a4437f834c"} Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.209800 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:17 crc kubenswrapper[4754]: W1011 04:23:17.220171 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9f26c36_2ab7_4b37_b3bd_c6ef1667c728.slice/crio-cc25c40fbec58c31d7d7f00857408d48582298a2d2583cb3961390b4f3414726 WatchSource:0}: Error finding container cc25c40fbec58c31d7d7f00857408d48582298a2d2583cb3961390b4f3414726: Status 404 returned error can't find the container with id cc25c40fbec58c31d7d7f00857408d48582298a2d2583cb3961390b4f3414726 Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.972761 4754 generic.go:334] "Generic (PLEG): container finished" podID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerID="265bfbce43848ce3a2abb8673c26ca2cc01fa7ff5038eb96bd2fb0af14635b7c" exitCode=0 Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.972908 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerDied","Data":"265bfbce43848ce3a2abb8673c26ca2cc01fa7ff5038eb96bd2fb0af14635b7c"} Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.973296 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerStarted","Data":"cc25c40fbec58c31d7d7f00857408d48582298a2d2583cb3961390b4f3414726"} Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.976012 4754 generic.go:334] "Generic (PLEG): container finished" podID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerID="9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb" exitCode=0 Oct 11 04:23:17 crc kubenswrapper[4754]: I1011 04:23:17.976148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerDied","Data":"9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb"} Oct 11 04:23:18 crc kubenswrapper[4754]: I1011 04:23:18.986290 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerStarted","Data":"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6"} Oct 11 04:23:18 crc kubenswrapper[4754]: I1011 04:23:18.991070 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerStarted","Data":"081d7f9c82b8fcac9539ced1d446390528d03b08ec61e1143d5246e04f60c300"} Oct 11 04:23:18 crc kubenswrapper[4754]: I1011 04:23:18.993855 4754 generic.go:334] "Generic (PLEG): container finished" podID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerID="527bc5e4bb7229fcce4dafe1f89b56d1663c5ed290d23cf8a90439a4437f834c" exitCode=0 Oct 11 04:23:18 crc kubenswrapper[4754]: I1011 04:23:18.993903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerDied","Data":"527bc5e4bb7229fcce4dafe1f89b56d1663c5ed290d23cf8a90439a4437f834c"} Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.007196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerStarted","Data":"ff831317a5273247e1c5674d0cc78b6644e102cf892867dd12bac77595869570"} Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.009998 4754 generic.go:334] "Generic (PLEG): container finished" podID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerID="8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6" exitCode=0 Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.010079 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerDied","Data":"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6"} Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.013297 4754 generic.go:334] "Generic (PLEG): container finished" podID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerID="081d7f9c82b8fcac9539ced1d446390528d03b08ec61e1143d5246e04f60c300" exitCode=0 Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.013334 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerDied","Data":"081d7f9c82b8fcac9539ced1d446390528d03b08ec61e1143d5246e04f60c300"} Oct 11 04:23:20 crc kubenswrapper[4754]: I1011 04:23:20.041272 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8w7nm" podStartSLOduration=2.560980044 podStartE2EDuration="7.04125052s" podCreationTimestamp="2025-10-11 04:23:13 +0000 UTC" firstStartedPulling="2025-10-11 04:23:14.928602238 +0000 UTC m=+4642.487547023" lastFinishedPulling="2025-10-11 04:23:19.408872714 +0000 UTC m=+4646.967817499" observedRunningTime="2025-10-11 04:23:20.036117586 +0000 UTC m=+4647.595062371" watchObservedRunningTime="2025-10-11 04:23:20.04125052 +0000 UTC m=+4647.600195305" Oct 11 04:23:21 crc kubenswrapper[4754]: I1011 04:23:21.059123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerStarted","Data":"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee"} Oct 11 04:23:21 crc kubenswrapper[4754]: I1011 04:23:21.062095 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerStarted","Data":"26296010c5f0569b49bfa4995bcd40f76991f9fdbbba4dc6ff1575b354a786eb"} Oct 11 04:23:21 crc kubenswrapper[4754]: I1011 04:23:21.087260 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:23:21 crc kubenswrapper[4754]: E1011 04:23:21.088029 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:23:21 crc kubenswrapper[4754]: I1011 04:23:21.102786 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vk9r2" podStartSLOduration=3.579516317 podStartE2EDuration="6.102762078s" podCreationTimestamp="2025-10-11 04:23:15 +0000 UTC" firstStartedPulling="2025-10-11 04:23:17.97880554 +0000 UTC m=+4645.537750365" lastFinishedPulling="2025-10-11 04:23:20.502051331 +0000 UTC m=+4648.060996126" observedRunningTime="2025-10-11 04:23:21.090344399 +0000 UTC m=+4648.649289194" watchObservedRunningTime="2025-10-11 04:23:21.102762078 +0000 UTC m=+4648.661706863" Oct 11 04:23:21 crc kubenswrapper[4754]: I1011 04:23:21.120084 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p6sh7" podStartSLOduration=2.682434981 podStartE2EDuration="5.120064244s" podCreationTimestamp="2025-10-11 04:23:16 +0000 UTC" firstStartedPulling="2025-10-11 04:23:17.977473133 +0000 UTC m=+4645.536417938" lastFinishedPulling="2025-10-11 04:23:20.415102416 +0000 UTC m=+4647.974047201" observedRunningTime="2025-10-11 04:23:21.112242104 +0000 UTC m=+4648.671186889" watchObservedRunningTime="2025-10-11 04:23:21.120064244 +0000 UTC m=+4648.679009029" Oct 11 04:23:23 crc kubenswrapper[4754]: I1011 04:23:23.970410 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:23 crc kubenswrapper[4754]: I1011 04:23:23.970861 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:24 crc kubenswrapper[4754]: I1011 04:23:24.024296 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:24 crc kubenswrapper[4754]: I1011 04:23:24.131802 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:25 crc kubenswrapper[4754]: I1011 04:23:25.811293 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:25 crc kubenswrapper[4754]: I1011 04:23:25.811690 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:25 crc kubenswrapper[4754]: I1011 04:23:25.860874 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.146537 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.617704 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.618105 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8w7nm" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="registry-server" containerID="cri-o://ff831317a5273247e1c5674d0cc78b6644e102cf892867dd12bac77595869570" gracePeriod=2 Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.647205 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.647254 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:26 crc kubenswrapper[4754]: I1011 04:23:26.696896 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.114179 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.114419 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerDied","Data":"ff831317a5273247e1c5674d0cc78b6644e102cf892867dd12bac77595869570"} Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.114383 4754 generic.go:334] "Generic (PLEG): container finished" podID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerID="ff831317a5273247e1c5674d0cc78b6644e102cf892867dd12bac77595869570" exitCode=0 Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.114532 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8w7nm" event={"ID":"a35811cf-3d9d-4eeb-8a59-65637cb443bb","Type":"ContainerDied","Data":"73bfaccfa39be9d811163eb095eb84d9749e39abd79cc4678c32b7f3c31681cb"} Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.114586 4754 scope.go:117] "RemoveContainer" containerID="ff831317a5273247e1c5674d0cc78b6644e102cf892867dd12bac77595869570" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.148183 4754 scope.go:117] "RemoveContainer" containerID="527bc5e4bb7229fcce4dafe1f89b56d1663c5ed290d23cf8a90439a4437f834c" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.169412 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.177649 4754 scope.go:117] "RemoveContainer" containerID="446efa27d7357c28d432336855f4e1ed04c4c19a6071e6869913db5183a3c940" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.280659 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities\") pod \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.280740 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncbgx\" (UniqueName: \"kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx\") pod \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.280762 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content\") pod \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\" (UID: \"a35811cf-3d9d-4eeb-8a59-65637cb443bb\") " Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.282407 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities" (OuterVolumeSpecName: "utilities") pod "a35811cf-3d9d-4eeb-8a59-65637cb443bb" (UID: "a35811cf-3d9d-4eeb-8a59-65637cb443bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.330738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a35811cf-3d9d-4eeb-8a59-65637cb443bb" (UID: "a35811cf-3d9d-4eeb-8a59-65637cb443bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.382815 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.382848 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35811cf-3d9d-4eeb-8a59-65637cb443bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.872605 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx" (OuterVolumeSpecName: "kube-api-access-ncbgx") pod "a35811cf-3d9d-4eeb-8a59-65637cb443bb" (UID: "a35811cf-3d9d-4eeb-8a59-65637cb443bb"). InnerVolumeSpecName "kube-api-access-ncbgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:23:27 crc kubenswrapper[4754]: I1011 04:23:27.892339 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncbgx\" (UniqueName: \"kubernetes.io/projected/a35811cf-3d9d-4eeb-8a59-65637cb443bb-kube-api-access-ncbgx\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.122042 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8w7nm" Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.151956 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.159666 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8w7nm"] Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.412919 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.413201 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vk9r2" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="registry-server" containerID="cri-o://2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee" gracePeriod=2 Oct 11 04:23:28 crc kubenswrapper[4754]: I1011 04:23:28.881178 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.014306 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities\") pod \"49e4300c-92a2-457c-86f4-7375802ac3f3\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.014417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content\") pod \"49e4300c-92a2-457c-86f4-7375802ac3f3\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.014475 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xggds\" (UniqueName: \"kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds\") pod \"49e4300c-92a2-457c-86f4-7375802ac3f3\" (UID: \"49e4300c-92a2-457c-86f4-7375802ac3f3\") " Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.015326 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities" (OuterVolumeSpecName: "utilities") pod "49e4300c-92a2-457c-86f4-7375802ac3f3" (UID: "49e4300c-92a2-457c-86f4-7375802ac3f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.022198 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds" (OuterVolumeSpecName: "kube-api-access-xggds") pod "49e4300c-92a2-457c-86f4-7375802ac3f3" (UID: "49e4300c-92a2-457c-86f4-7375802ac3f3"). InnerVolumeSpecName "kube-api-access-xggds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.099937 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49e4300c-92a2-457c-86f4-7375802ac3f3" (UID: "49e4300c-92a2-457c-86f4-7375802ac3f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.113484 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" path="/var/lib/kubelet/pods/a35811cf-3d9d-4eeb-8a59-65637cb443bb/volumes" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.119639 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.119678 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e4300c-92a2-457c-86f4-7375802ac3f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.119690 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xggds\" (UniqueName: \"kubernetes.io/projected/49e4300c-92a2-457c-86f4-7375802ac3f3-kube-api-access-xggds\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.135291 4754 generic.go:334] "Generic (PLEG): container finished" podID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerID="2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee" exitCode=0 Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.135337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerDied","Data":"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee"} Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.135371 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk9r2" event={"ID":"49e4300c-92a2-457c-86f4-7375802ac3f3","Type":"ContainerDied","Data":"abcb3d1738cc69ee2b12ddb6f2ccd2ea02380c139a0bdaf3fda8b6ac01ea0dc7"} Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.135391 4754 scope.go:117] "RemoveContainer" containerID="2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.135396 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk9r2" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.159432 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.162699 4754 scope.go:117] "RemoveContainer" containerID="8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.170645 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vk9r2"] Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.690116 4754 scope.go:117] "RemoveContainer" containerID="9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.751203 4754 scope.go:117] "RemoveContainer" containerID="2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee" Oct 11 04:23:29 crc kubenswrapper[4754]: E1011 04:23:29.751630 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee\": container with ID starting with 2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee not found: ID does not exist" containerID="2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.751670 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee"} err="failed to get container status \"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee\": rpc error: code = NotFound desc = could not find container \"2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee\": container with ID starting with 2ed4b15310640f9786ea7a84a62d427da06fdda4da2680588da304bd34b873ee not found: ID does not exist" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.751697 4754 scope.go:117] "RemoveContainer" containerID="8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6" Oct 11 04:23:29 crc kubenswrapper[4754]: E1011 04:23:29.752170 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6\": container with ID starting with 8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6 not found: ID does not exist" containerID="8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.752234 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6"} err="failed to get container status \"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6\": rpc error: code = NotFound desc = could not find container \"8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6\": container with ID starting with 8e4c0abf5364132cd38a7012a3eb8e3057d0c4785265e9a6265571a783084ab6 not found: ID does not exist" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.752267 4754 scope.go:117] "RemoveContainer" containerID="9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb" Oct 11 04:23:29 crc kubenswrapper[4754]: E1011 04:23:29.752616 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb\": container with ID starting with 9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb not found: ID does not exist" containerID="9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb" Oct 11 04:23:29 crc kubenswrapper[4754]: I1011 04:23:29.752651 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb"} err="failed to get container status \"9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb\": rpc error: code = NotFound desc = could not find container \"9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb\": container with ID starting with 9ca250759c39b2468554f92d1ba8e427fae94b454077ab0c1b763bc7095788bb not found: ID does not exist" Oct 11 04:23:30 crc kubenswrapper[4754]: I1011 04:23:30.817440 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:30 crc kubenswrapper[4754]: I1011 04:23:30.817778 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p6sh7" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="registry-server" containerID="cri-o://26296010c5f0569b49bfa4995bcd40f76991f9fdbbba4dc6ff1575b354a786eb" gracePeriod=2 Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.103749 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" path="/var/lib/kubelet/pods/49e4300c-92a2-457c-86f4-7375802ac3f3/volumes" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.166597 4754 generic.go:334] "Generic (PLEG): container finished" podID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerID="26296010c5f0569b49bfa4995bcd40f76991f9fdbbba4dc6ff1575b354a786eb" exitCode=0 Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.166643 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerDied","Data":"26296010c5f0569b49bfa4995bcd40f76991f9fdbbba4dc6ff1575b354a786eb"} Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.279340 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.469584 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content\") pod \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.469668 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities\") pod \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.469692 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb2cs\" (UniqueName: \"kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs\") pod \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\" (UID: \"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728\") " Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.470628 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities" (OuterVolumeSpecName: "utilities") pod "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" (UID: "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.477177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs" (OuterVolumeSpecName: "kube-api-access-lb2cs") pod "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" (UID: "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728"). InnerVolumeSpecName "kube-api-access-lb2cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.515361 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" (UID: "d9f26c36-2ab7-4b37-b3bd-c6ef1667c728"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.572519 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.572571 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:31 crc kubenswrapper[4754]: I1011 04:23:31.572585 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb2cs\" (UniqueName: \"kubernetes.io/projected/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728-kube-api-access-lb2cs\") on node \"crc\" DevicePath \"\"" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.083650 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:23:32 crc kubenswrapper[4754]: E1011 04:23:32.084559 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.179657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6sh7" event={"ID":"d9f26c36-2ab7-4b37-b3bd-c6ef1667c728","Type":"ContainerDied","Data":"cc25c40fbec58c31d7d7f00857408d48582298a2d2583cb3961390b4f3414726"} Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.179750 4754 scope.go:117] "RemoveContainer" containerID="26296010c5f0569b49bfa4995bcd40f76991f9fdbbba4dc6ff1575b354a786eb" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.179819 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6sh7" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.200359 4754 scope.go:117] "RemoveContainer" containerID="081d7f9c82b8fcac9539ced1d446390528d03b08ec61e1143d5246e04f60c300" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.221349 4754 scope.go:117] "RemoveContainer" containerID="265bfbce43848ce3a2abb8673c26ca2cc01fa7ff5038eb96bd2fb0af14635b7c" Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.294780 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:32 crc kubenswrapper[4754]: I1011 04:23:32.303411 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p6sh7"] Oct 11 04:23:33 crc kubenswrapper[4754]: I1011 04:23:33.101765 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" path="/var/lib/kubelet/pods/d9f26c36-2ab7-4b37-b3bd-c6ef1667c728/volumes" Oct 11 04:23:43 crc kubenswrapper[4754]: I1011 04:23:43.090559 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:23:43 crc kubenswrapper[4754]: E1011 04:23:43.091339 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:23:57 crc kubenswrapper[4754]: I1011 04:23:57.083589 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:23:57 crc kubenswrapper[4754]: E1011 04:23:57.084361 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:24:12 crc kubenswrapper[4754]: I1011 04:24:12.088564 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:24:12 crc kubenswrapper[4754]: E1011 04:24:12.090420 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:24:27 crc kubenswrapper[4754]: I1011 04:24:27.083688 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:24:27 crc kubenswrapper[4754]: E1011 04:24:27.084725 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:24:42 crc kubenswrapper[4754]: I1011 04:24:42.084468 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:24:42 crc kubenswrapper[4754]: E1011 04:24:42.085353 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:24:53 crc kubenswrapper[4754]: I1011 04:24:53.098950 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:24:53 crc kubenswrapper[4754]: E1011 04:24:53.100113 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:25:08 crc kubenswrapper[4754]: I1011 04:25:08.083827 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:25:08 crc kubenswrapper[4754]: E1011 04:25:08.084606 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:25:22 crc kubenswrapper[4754]: I1011 04:25:22.084358 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:25:22 crc kubenswrapper[4754]: E1011 04:25:22.085354 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:25:33 crc kubenswrapper[4754]: I1011 04:25:33.091261 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:25:33 crc kubenswrapper[4754]: E1011 04:25:33.093388 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:25:48 crc kubenswrapper[4754]: I1011 04:25:48.084276 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:25:48 crc kubenswrapper[4754]: E1011 04:25:48.085196 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:26:03 crc kubenswrapper[4754]: I1011 04:26:03.089645 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:26:03 crc kubenswrapper[4754]: E1011 04:26:03.090332 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:26:15 crc kubenswrapper[4754]: I1011 04:26:15.083645 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:26:15 crc kubenswrapper[4754]: E1011 04:26:15.084506 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:26:30 crc kubenswrapper[4754]: I1011 04:26:30.083989 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:26:30 crc kubenswrapper[4754]: E1011 04:26:30.085740 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:26:41 crc kubenswrapper[4754]: I1011 04:26:41.086353 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:26:42 crc kubenswrapper[4754]: I1011 04:26:42.001108 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486"} Oct 11 04:28:40 crc kubenswrapper[4754]: I1011 04:28:40.196995 4754 generic.go:334] "Generic (PLEG): container finished" podID="1b1a049c-cb14-4ec6-8843-69950502bce7" containerID="e0f90349f5516615b638747af33a2d943849a0bc18247d391e40f0802afb7754" exitCode=1 Oct 11 04:28:40 crc kubenswrapper[4754]: I1011 04:28:40.198733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1b1a049c-cb14-4ec6-8843-69950502bce7","Type":"ContainerDied","Data":"e0f90349f5516615b638747af33a2d943849a0bc18247d391e40f0802afb7754"} Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.610187 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.761273 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.761405 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.761444 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n2b5\" (UniqueName: \"kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.761575 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762366 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762523 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data" (OuterVolumeSpecName: "config-data") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762572 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762657 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762713 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.762765 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1b1a049c-cb14-4ec6-8843-69950502bce7\" (UID: \"1b1a049c-cb14-4ec6-8843-69950502bce7\") " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.763262 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.764426 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.764453 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.767451 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.769626 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.772636 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5" (OuterVolumeSpecName: "kube-api-access-7n2b5") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "kube-api-access-7n2b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.793426 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.800816 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.816573 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.818620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1b1a049c-cb14-4ec6-8843-69950502bce7" (UID: "1b1a049c-cb14-4ec6-8843-69950502bce7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866581 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866616 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866756 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n2b5\" (UniqueName: \"kubernetes.io/projected/1b1a049c-cb14-4ec6-8843-69950502bce7-kube-api-access-7n2b5\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866771 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1b1a049c-cb14-4ec6-8843-69950502bce7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866780 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866793 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.866801 4754 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1b1a049c-cb14-4ec6-8843-69950502bce7-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.901722 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 11 04:28:41 crc kubenswrapper[4754]: I1011 04:28:41.968311 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:42 crc kubenswrapper[4754]: I1011 04:28:42.218337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1b1a049c-cb14-4ec6-8843-69950502bce7","Type":"ContainerDied","Data":"4ebc0e884da73328a56be46c539189af12ccb1f7325c7ada38130d3ac42764eb"} Oct 11 04:28:42 crc kubenswrapper[4754]: I1011 04:28:42.218390 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ebc0e884da73328a56be46c539189af12ccb1f7325c7ada38130d3ac42764eb" Oct 11 04:28:42 crc kubenswrapper[4754]: I1011 04:28:42.218467 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.161761 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162754 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162773 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162783 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162789 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162809 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162818 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162831 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162838 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162853 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162861 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162880 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162888 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162901 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1a049c-cb14-4ec6-8843-69950502bce7" containerName="tempest-tests-tempest-tests-runner" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162908 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1a049c-cb14-4ec6-8843-69950502bce7" containerName="tempest-tests-tempest-tests-runner" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162939 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162946 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="extract-content" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.162980 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.162989 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="extract-utilities" Oct 11 04:28:44 crc kubenswrapper[4754]: E1011 04:28:44.163002 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.163010 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.163256 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e4300c-92a2-457c-86f4-7375802ac3f3" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.163273 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1a049c-cb14-4ec6-8843-69950502bce7" containerName="tempest-tests-tempest-tests-runner" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.163300 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f26c36-2ab7-4b37-b3bd-c6ef1667c728" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.163315 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a35811cf-3d9d-4eeb-8a59-65637cb443bb" containerName="registry-server" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.164096 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.164195 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.166764 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdw6z" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.323746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.323872 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dv6\" (UniqueName: \"kubernetes.io/projected/8d37a1cd-4bdf-4c04-9347-40fce1afe82f-kube-api-access-c4dv6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.425415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dv6\" (UniqueName: \"kubernetes.io/projected/8d37a1cd-4bdf-4c04-9347-40fce1afe82f-kube-api-access-c4dv6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.425559 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.426056 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.447159 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dv6\" (UniqueName: \"kubernetes.io/projected/8d37a1cd-4bdf-4c04-9347-40fce1afe82f-kube-api-access-c4dv6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.454143 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8d37a1cd-4bdf-4c04-9347-40fce1afe82f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.505465 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.952711 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 11 04:28:44 crc kubenswrapper[4754]: I1011 04:28:44.970140 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:28:45 crc kubenswrapper[4754]: I1011 04:28:45.243139 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8d37a1cd-4bdf-4c04-9347-40fce1afe82f","Type":"ContainerStarted","Data":"3cf83c8de13da1c4b1350aeca51ea8ed15c98ae1b4ffdb62b163e77722a68792"} Oct 11 04:28:46 crc kubenswrapper[4754]: I1011 04:28:46.254574 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8d37a1cd-4bdf-4c04-9347-40fce1afe82f","Type":"ContainerStarted","Data":"e8dcbb0930a27b482d2f6e3cb70dbbe298536861d62c1656196e65ec1d6a05e5"} Oct 11 04:28:46 crc kubenswrapper[4754]: I1011 04:28:46.282876 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.453491166 podStartE2EDuration="2.282807369s" podCreationTimestamp="2025-10-11 04:28:44 +0000 UTC" firstStartedPulling="2025-10-11 04:28:44.969746038 +0000 UTC m=+4972.528690823" lastFinishedPulling="2025-10-11 04:28:45.799062241 +0000 UTC m=+4973.358007026" observedRunningTime="2025-10-11 04:28:46.265628706 +0000 UTC m=+4973.824573531" watchObservedRunningTime="2025-10-11 04:28:46.282807369 +0000 UTC m=+4973.841752194" Oct 11 04:29:00 crc kubenswrapper[4754]: I1011 04:29:00.736957 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:29:00 crc kubenswrapper[4754]: I1011 04:29:00.737955 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:29:30 crc kubenswrapper[4754]: I1011 04:29:30.736956 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:29:30 crc kubenswrapper[4754]: I1011 04:29:30.738214 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.132744 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xbq7q/must-gather-lfcgj"] Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.134676 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.136146 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xbq7q"/"default-dockercfg-2fhsj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.137412 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xbq7q"/"openshift-service-ca.crt" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.139315 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xbq7q"/"kube-root-ca.crt" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.156561 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xbq7q/must-gather-lfcgj"] Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.276131 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.276186 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4ct\" (UniqueName: \"kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.378592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.378653 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4ct\" (UniqueName: \"kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.379359 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.397736 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4ct\" (UniqueName: \"kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct\") pod \"must-gather-lfcgj\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.453373 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:29:32 crc kubenswrapper[4754]: I1011 04:29:32.899220 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xbq7q/must-gather-lfcgj"] Oct 11 04:29:33 crc kubenswrapper[4754]: I1011 04:29:33.745132 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" event={"ID":"da1fb582-ac79-4d1a-8380-2feb95343104","Type":"ContainerStarted","Data":"d0ca40fe6c531e08e546511986bb8f96eb2a5818d0326c9ebaa430e776641c60"} Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.159534 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.162358 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.173062 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.323032 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7j6q\" (UniqueName: \"kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.323123 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.323465 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.425053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.425149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7j6q\" (UniqueName: \"kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.425208 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.425568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.425646 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.465236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7j6q\" (UniqueName: \"kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q\") pod \"redhat-marketplace-hjcnk\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:38 crc kubenswrapper[4754]: I1011 04:29:38.524949 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.174760 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:39 crc kubenswrapper[4754]: W1011 04:29:39.181753 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8269e686_4b7f_4813_96ac_b2525c1a6f7f.slice/crio-d46727694b4a874282cf369c119bde9b9e7fe2fd4a68dd5b753a49fb34c7a45b WatchSource:0}: Error finding container d46727694b4a874282cf369c119bde9b9e7fe2fd4a68dd5b753a49fb34c7a45b: Status 404 returned error can't find the container with id d46727694b4a874282cf369c119bde9b9e7fe2fd4a68dd5b753a49fb34c7a45b Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.818040 4754 generic.go:334] "Generic (PLEG): container finished" podID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerID="9735a938e4ea6c7072e2995af0b3d6e74765d01ad81f68a065647593f692bd8a" exitCode=0 Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.818359 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerDied","Data":"9735a938e4ea6c7072e2995af0b3d6e74765d01ad81f68a065647593f692bd8a"} Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.818387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerStarted","Data":"d46727694b4a874282cf369c119bde9b9e7fe2fd4a68dd5b753a49fb34c7a45b"} Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.824341 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" event={"ID":"da1fb582-ac79-4d1a-8380-2feb95343104","Type":"ContainerStarted","Data":"e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5"} Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.824481 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" event={"ID":"da1fb582-ac79-4d1a-8380-2feb95343104","Type":"ContainerStarted","Data":"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569"} Oct 11 04:29:39 crc kubenswrapper[4754]: I1011 04:29:39.860578 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" podStartSLOduration=2.030354697 podStartE2EDuration="7.860559525s" podCreationTimestamp="2025-10-11 04:29:32 +0000 UTC" firstStartedPulling="2025-10-11 04:29:32.90645665 +0000 UTC m=+5020.465401435" lastFinishedPulling="2025-10-11 04:29:38.736661478 +0000 UTC m=+5026.295606263" observedRunningTime="2025-10-11 04:29:39.8525613 +0000 UTC m=+5027.411506085" watchObservedRunningTime="2025-10-11 04:29:39.860559525 +0000 UTC m=+5027.419504310" Oct 11 04:29:41 crc kubenswrapper[4754]: I1011 04:29:41.858545 4754 generic.go:334] "Generic (PLEG): container finished" podID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerID="8a46ac57335855e40f3893411db2b23b8c015e13b6aed8edab4439e8b0c2aa16" exitCode=0 Oct 11 04:29:41 crc kubenswrapper[4754]: I1011 04:29:41.861597 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerDied","Data":"8a46ac57335855e40f3893411db2b23b8c015e13b6aed8edab4439e8b0c2aa16"} Oct 11 04:29:42 crc kubenswrapper[4754]: I1011 04:29:42.884394 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerStarted","Data":"849f59d186d5202f7db0b5e7b14d803ded6d4bb6995179540c6ea2edbbebe568"} Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.704709 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-5zgtw"] Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.706285 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.836452 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmkqb\" (UniqueName: \"kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.836787 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.912368 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hjcnk" podStartSLOduration=3.100522663 podStartE2EDuration="5.912346891s" podCreationTimestamp="2025-10-11 04:29:38 +0000 UTC" firstStartedPulling="2025-10-11 04:29:39.81981188 +0000 UTC m=+5027.378756665" lastFinishedPulling="2025-10-11 04:29:42.631636098 +0000 UTC m=+5030.190580893" observedRunningTime="2025-10-11 04:29:43.908801001 +0000 UTC m=+5031.467745796" watchObservedRunningTime="2025-10-11 04:29:43.912346891 +0000 UTC m=+5031.471291686" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.939293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmkqb\" (UniqueName: \"kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.939646 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.939774 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:43 crc kubenswrapper[4754]: I1011 04:29:43.956755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmkqb\" (UniqueName: \"kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb\") pod \"crc-debug-5zgtw\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:44 crc kubenswrapper[4754]: I1011 04:29:44.026060 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:29:44 crc kubenswrapper[4754]: I1011 04:29:44.903947 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" event={"ID":"e759210a-d281-49b1-80af-a9d85d517f03","Type":"ContainerStarted","Data":"e89554290009d975d818965eac8f421bc97ed73b3a59e6c05dfc082ea8e2a463"} Oct 11 04:29:45 crc kubenswrapper[4754]: E1011 04:29:45.619579 4754 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.188:50564->38.102.83.188:46049: write tcp 38.102.83.188:50564->38.102.83.188:46049: write: broken pipe Oct 11 04:29:48 crc kubenswrapper[4754]: I1011 04:29:48.526001 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:48 crc kubenswrapper[4754]: I1011 04:29:48.526574 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:48 crc kubenswrapper[4754]: I1011 04:29:48.581820 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:49 crc kubenswrapper[4754]: I1011 04:29:49.018457 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:49 crc kubenswrapper[4754]: I1011 04:29:49.068674 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:50 crc kubenswrapper[4754]: I1011 04:29:50.958892 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hjcnk" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="registry-server" containerID="cri-o://849f59d186d5202f7db0b5e7b14d803ded6d4bb6995179540c6ea2edbbebe568" gracePeriod=2 Oct 11 04:29:51 crc kubenswrapper[4754]: I1011 04:29:51.971004 4754 generic.go:334] "Generic (PLEG): container finished" podID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerID="849f59d186d5202f7db0b5e7b14d803ded6d4bb6995179540c6ea2edbbebe568" exitCode=0 Oct 11 04:29:51 crc kubenswrapper[4754]: I1011 04:29:51.971074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerDied","Data":"849f59d186d5202f7db0b5e7b14d803ded6d4bb6995179540c6ea2edbbebe568"} Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.644339 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.769659 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content\") pod \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.769902 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7j6q\" (UniqueName: \"kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q\") pod \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.769981 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities\") pod \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\" (UID: \"8269e686-4b7f-4813-96ac-b2525c1a6f7f\") " Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.770917 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities" (OuterVolumeSpecName: "utilities") pod "8269e686-4b7f-4813-96ac-b2525c1a6f7f" (UID: "8269e686-4b7f-4813-96ac-b2525c1a6f7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.777508 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q" (OuterVolumeSpecName: "kube-api-access-x7j6q") pod "8269e686-4b7f-4813-96ac-b2525c1a6f7f" (UID: "8269e686-4b7f-4813-96ac-b2525c1a6f7f"). InnerVolumeSpecName "kube-api-access-x7j6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.786321 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8269e686-4b7f-4813-96ac-b2525c1a6f7f" (UID: "8269e686-4b7f-4813-96ac-b2525c1a6f7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.874089 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7j6q\" (UniqueName: \"kubernetes.io/projected/8269e686-4b7f-4813-96ac-b2525c1a6f7f-kube-api-access-x7j6q\") on node \"crc\" DevicePath \"\"" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.874144 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:29:54 crc kubenswrapper[4754]: I1011 04:29:54.874158 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8269e686-4b7f-4813-96ac-b2525c1a6f7f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.000203 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjcnk" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.000406 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjcnk" event={"ID":"8269e686-4b7f-4813-96ac-b2525c1a6f7f","Type":"ContainerDied","Data":"d46727694b4a874282cf369c119bde9b9e7fe2fd4a68dd5b753a49fb34c7a45b"} Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.000481 4754 scope.go:117] "RemoveContainer" containerID="849f59d186d5202f7db0b5e7b14d803ded6d4bb6995179540c6ea2edbbebe568" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.002616 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" event={"ID":"e759210a-d281-49b1-80af-a9d85d517f03","Type":"ContainerStarted","Data":"7ea42daf2fa2d1e00dcc12d0ad3998119037f69aa443bf0bb404fac44838f6cc"} Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.025523 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" podStartSLOduration=1.7276339200000002 podStartE2EDuration="12.025503625s" podCreationTimestamp="2025-10-11 04:29:43 +0000 UTC" firstStartedPulling="2025-10-11 04:29:44.06293403 +0000 UTC m=+5031.621878815" lastFinishedPulling="2025-10-11 04:29:54.360803725 +0000 UTC m=+5041.919748520" observedRunningTime="2025-10-11 04:29:55.020531275 +0000 UTC m=+5042.579476060" watchObservedRunningTime="2025-10-11 04:29:55.025503625 +0000 UTC m=+5042.584448410" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.028617 4754 scope.go:117] "RemoveContainer" containerID="8a46ac57335855e40f3893411db2b23b8c015e13b6aed8edab4439e8b0c2aa16" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.054330 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.062352 4754 scope.go:117] "RemoveContainer" containerID="9735a938e4ea6c7072e2995af0b3d6e74765d01ad81f68a065647593f692bd8a" Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.072573 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjcnk"] Oct 11 04:29:55 crc kubenswrapper[4754]: I1011 04:29:55.118283 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" path="/var/lib/kubelet/pods/8269e686-4b7f-4813-96ac-b2525c1a6f7f/volumes" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.145526 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8"] Oct 11 04:30:00 crc kubenswrapper[4754]: E1011 04:30:00.146691 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.146709 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4754]: E1011 04:30:00.146754 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.146763 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4754]: E1011 04:30:00.146791 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.146802 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.147076 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8269e686-4b7f-4813-96ac-b2525c1a6f7f" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.148084 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.150574 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.151265 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.161005 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8"] Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.290149 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.290558 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.291117 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p244b\" (UniqueName: \"kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.393910 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p244b\" (UniqueName: \"kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.394154 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.394256 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.395452 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.412355 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.418410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p244b\" (UniqueName: \"kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b\") pod \"collect-profiles-29335950-lgxt8\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.468360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.737271 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.737635 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.737679 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.738586 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:30:00 crc kubenswrapper[4754]: I1011 04:30:00.742606 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486" gracePeriod=600 Oct 11 04:30:01 crc kubenswrapper[4754]: I1011 04:30:01.041606 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8"] Oct 11 04:30:01 crc kubenswrapper[4754]: I1011 04:30:01.062687 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" event={"ID":"2f74499c-5dcc-4e05-ab24-7e966bd10711","Type":"ContainerStarted","Data":"7a47175394bf3d8134d188a18b2fce54bf1f87da00d8d2be9bd0ce72952f3c91"} Oct 11 04:30:01 crc kubenswrapper[4754]: I1011 04:30:01.069169 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486" exitCode=0 Oct 11 04:30:01 crc kubenswrapper[4754]: I1011 04:30:01.069248 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486"} Oct 11 04:30:01 crc kubenswrapper[4754]: I1011 04:30:01.069406 4754 scope.go:117] "RemoveContainer" containerID="c0b97de9e1e69550ece9671c1646d7c24cfc5ebcda36b67abc5467070afbf1cc" Oct 11 04:30:02 crc kubenswrapper[4754]: I1011 04:30:02.091183 4754 generic.go:334] "Generic (PLEG): container finished" podID="2f74499c-5dcc-4e05-ab24-7e966bd10711" containerID="93d553c148218e4921e886d84e455e8dcf7511a1fed27624000cbed8b5fa51f4" exitCode=0 Oct 11 04:30:02 crc kubenswrapper[4754]: I1011 04:30:02.091854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" event={"ID":"2f74499c-5dcc-4e05-ab24-7e966bd10711","Type":"ContainerDied","Data":"93d553c148218e4921e886d84e455e8dcf7511a1fed27624000cbed8b5fa51f4"} Oct 11 04:30:02 crc kubenswrapper[4754]: I1011 04:30:02.095885 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerStarted","Data":"6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038"} Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.453784 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.562689 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume\") pod \"2f74499c-5dcc-4e05-ab24-7e966bd10711\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.563288 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p244b\" (UniqueName: \"kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b\") pod \"2f74499c-5dcc-4e05-ab24-7e966bd10711\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.563350 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume\") pod \"2f74499c-5dcc-4e05-ab24-7e966bd10711\" (UID: \"2f74499c-5dcc-4e05-ab24-7e966bd10711\") " Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.564411 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f74499c-5dcc-4e05-ab24-7e966bd10711" (UID: "2f74499c-5dcc-4e05-ab24-7e966bd10711"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.571883 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f74499c-5dcc-4e05-ab24-7e966bd10711" (UID: "2f74499c-5dcc-4e05-ab24-7e966bd10711"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.573397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b" (OuterVolumeSpecName: "kube-api-access-p244b") pod "2f74499c-5dcc-4e05-ab24-7e966bd10711" (UID: "2f74499c-5dcc-4e05-ab24-7e966bd10711"). InnerVolumeSpecName "kube-api-access-p244b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.666221 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f74499c-5dcc-4e05-ab24-7e966bd10711-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.666262 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p244b\" (UniqueName: \"kubernetes.io/projected/2f74499c-5dcc-4e05-ab24-7e966bd10711-kube-api-access-p244b\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:03 crc kubenswrapper[4754]: I1011 04:30:03.666273 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f74499c-5dcc-4e05-ab24-7e966bd10711-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:04 crc kubenswrapper[4754]: I1011 04:30:04.117106 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" event={"ID":"2f74499c-5dcc-4e05-ab24-7e966bd10711","Type":"ContainerDied","Data":"7a47175394bf3d8134d188a18b2fce54bf1f87da00d8d2be9bd0ce72952f3c91"} Oct 11 04:30:04 crc kubenswrapper[4754]: I1011 04:30:04.117585 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a47175394bf3d8134d188a18b2fce54bf1f87da00d8d2be9bd0ce72952f3c91" Oct 11 04:30:04 crc kubenswrapper[4754]: I1011 04:30:04.117315 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-lgxt8" Oct 11 04:30:04 crc kubenswrapper[4754]: I1011 04:30:04.528225 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm"] Oct 11 04:30:04 crc kubenswrapper[4754]: I1011 04:30:04.536934 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-ttmkm"] Oct 11 04:30:05 crc kubenswrapper[4754]: I1011 04:30:05.097035 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a09745f4-c2b2-45b0-98f8-23779ee4144e" path="/var/lib/kubelet/pods/a09745f4-c2b2-45b0-98f8-23779ee4144e/volumes" Oct 11 04:30:08 crc kubenswrapper[4754]: I1011 04:30:08.164284 4754 scope.go:117] "RemoveContainer" containerID="1307327f11d36ec0cd018b1643c7192cbbe028e37a1a1cfa6e0928d63de5c4c9" Oct 11 04:30:39 crc kubenswrapper[4754]: I1011 04:30:39.463105 4754 generic.go:334] "Generic (PLEG): container finished" podID="e759210a-d281-49b1-80af-a9d85d517f03" containerID="7ea42daf2fa2d1e00dcc12d0ad3998119037f69aa443bf0bb404fac44838f6cc" exitCode=0 Oct 11 04:30:39 crc kubenswrapper[4754]: I1011 04:30:39.463191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" event={"ID":"e759210a-d281-49b1-80af-a9d85d517f03","Type":"ContainerDied","Data":"7ea42daf2fa2d1e00dcc12d0ad3998119037f69aa443bf0bb404fac44838f6cc"} Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.579827 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.620624 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-5zgtw"] Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.630322 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-5zgtw"] Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.658627 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host\") pod \"e759210a-d281-49b1-80af-a9d85d517f03\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.658753 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmkqb\" (UniqueName: \"kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb\") pod \"e759210a-d281-49b1-80af-a9d85d517f03\" (UID: \"e759210a-d281-49b1-80af-a9d85d517f03\") " Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.658937 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host" (OuterVolumeSpecName: "host") pod "e759210a-d281-49b1-80af-a9d85d517f03" (UID: "e759210a-d281-49b1-80af-a9d85d517f03"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.659832 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e759210a-d281-49b1-80af-a9d85d517f03-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.680868 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb" (OuterVolumeSpecName: "kube-api-access-gmkqb") pod "e759210a-d281-49b1-80af-a9d85d517f03" (UID: "e759210a-d281-49b1-80af-a9d85d517f03"). InnerVolumeSpecName "kube-api-access-gmkqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:30:40 crc kubenswrapper[4754]: I1011 04:30:40.762557 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmkqb\" (UniqueName: \"kubernetes.io/projected/e759210a-d281-49b1-80af-a9d85d517f03-kube-api-access-gmkqb\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.095127 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e759210a-d281-49b1-80af-a9d85d517f03" path="/var/lib/kubelet/pods/e759210a-d281-49b1-80af-a9d85d517f03/volumes" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.496020 4754 scope.go:117] "RemoveContainer" containerID="7ea42daf2fa2d1e00dcc12d0ad3998119037f69aa443bf0bb404fac44838f6cc" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.496108 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-5zgtw" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.810210 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-65grl"] Oct 11 04:30:41 crc kubenswrapper[4754]: E1011 04:30:41.810700 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e759210a-d281-49b1-80af-a9d85d517f03" containerName="container-00" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.810719 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e759210a-d281-49b1-80af-a9d85d517f03" containerName="container-00" Oct 11 04:30:41 crc kubenswrapper[4754]: E1011 04:30:41.810730 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74499c-5dcc-4e05-ab24-7e966bd10711" containerName="collect-profiles" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.810739 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74499c-5dcc-4e05-ab24-7e966bd10711" containerName="collect-profiles" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.811033 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74499c-5dcc-4e05-ab24-7e966bd10711" containerName="collect-profiles" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.811065 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e759210a-d281-49b1-80af-a9d85d517f03" containerName="container-00" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.811852 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.888752 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf9c4\" (UniqueName: \"kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.889118 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.991467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.991541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf9c4\" (UniqueName: \"kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:41 crc kubenswrapper[4754]: I1011 04:30:41.991614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:42 crc kubenswrapper[4754]: I1011 04:30:42.014554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf9c4\" (UniqueName: \"kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4\") pod \"crc-debug-65grl\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:42 crc kubenswrapper[4754]: I1011 04:30:42.130690 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:42 crc kubenswrapper[4754]: I1011 04:30:42.506686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-65grl" event={"ID":"650f6681-180b-4498-bfb2-5c102b790fef","Type":"ContainerStarted","Data":"6f6475b18e6995c2aa8c39424da0dd87d301b8c7aebc7651b6dc942630879789"} Oct 11 04:30:42 crc kubenswrapper[4754]: I1011 04:30:42.507075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-65grl" event={"ID":"650f6681-180b-4498-bfb2-5c102b790fef","Type":"ContainerStarted","Data":"58e31e45311751e34af35b52f8d06688f7f5c426ba892e05c327164b5c9bb438"} Oct 11 04:30:42 crc kubenswrapper[4754]: I1011 04:30:42.519426 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xbq7q/crc-debug-65grl" podStartSLOduration=1.519401944 podStartE2EDuration="1.519401944s" podCreationTimestamp="2025-10-11 04:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:30:42.518698124 +0000 UTC m=+5090.077642939" watchObservedRunningTime="2025-10-11 04:30:42.519401944 +0000 UTC m=+5090.078346729" Oct 11 04:30:43 crc kubenswrapper[4754]: I1011 04:30:43.520122 4754 generic.go:334] "Generic (PLEG): container finished" podID="650f6681-180b-4498-bfb2-5c102b790fef" containerID="6f6475b18e6995c2aa8c39424da0dd87d301b8c7aebc7651b6dc942630879789" exitCode=0 Oct 11 04:30:43 crc kubenswrapper[4754]: I1011 04:30:43.520197 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-65grl" event={"ID":"650f6681-180b-4498-bfb2-5c102b790fef","Type":"ContainerDied","Data":"6f6475b18e6995c2aa8c39424da0dd87d301b8c7aebc7651b6dc942630879789"} Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.647851 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.746686 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host\") pod \"650f6681-180b-4498-bfb2-5c102b790fef\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.746785 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host" (OuterVolumeSpecName: "host") pod "650f6681-180b-4498-bfb2-5c102b790fef" (UID: "650f6681-180b-4498-bfb2-5c102b790fef"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.746925 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf9c4\" (UniqueName: \"kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4\") pod \"650f6681-180b-4498-bfb2-5c102b790fef\" (UID: \"650f6681-180b-4498-bfb2-5c102b790fef\") " Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.747432 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/650f6681-180b-4498-bfb2-5c102b790fef-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.753485 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4" (OuterVolumeSpecName: "kube-api-access-mf9c4") pod "650f6681-180b-4498-bfb2-5c102b790fef" (UID: "650f6681-180b-4498-bfb2-5c102b790fef"). InnerVolumeSpecName "kube-api-access-mf9c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:30:44 crc kubenswrapper[4754]: I1011 04:30:44.849026 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf9c4\" (UniqueName: \"kubernetes.io/projected/650f6681-180b-4498-bfb2-5c102b790fef-kube-api-access-mf9c4\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:45 crc kubenswrapper[4754]: I1011 04:30:45.372124 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-65grl"] Oct 11 04:30:45 crc kubenswrapper[4754]: I1011 04:30:45.383770 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-65grl"] Oct 11 04:30:45 crc kubenswrapper[4754]: I1011 04:30:45.544033 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58e31e45311751e34af35b52f8d06688f7f5c426ba892e05c327164b5c9bb438" Oct 11 04:30:45 crc kubenswrapper[4754]: I1011 04:30:45.544091 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-65grl" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.919808 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-xgqhb"] Oct 11 04:30:46 crc kubenswrapper[4754]: E1011 04:30:46.921027 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650f6681-180b-4498-bfb2-5c102b790fef" containerName="container-00" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.921051 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="650f6681-180b-4498-bfb2-5c102b790fef" containerName="container-00" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.921370 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="650f6681-180b-4498-bfb2-5c102b790fef" containerName="container-00" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.922430 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.998330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7n6f\" (UniqueName: \"kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:46 crc kubenswrapper[4754]: I1011 04:30:46.998389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.101923 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7n6f\" (UniqueName: \"kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.101998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.102238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.104523 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="650f6681-180b-4498-bfb2-5c102b790fef" path="/var/lib/kubelet/pods/650f6681-180b-4498-bfb2-5c102b790fef/volumes" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.128880 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7n6f\" (UniqueName: \"kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f\") pod \"crc-debug-xgqhb\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.244548 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.570677 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" event={"ID":"937fe477-f212-4b46-9a9a-f4392418ffdc","Type":"ContainerStarted","Data":"da4e17e52c4345d1d6b41fdf864fe0f50fbd8eb87b6a7eeecbc6666bc6797f4e"} Oct 11 04:30:47 crc kubenswrapper[4754]: I1011 04:30:47.757768 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6658bb6dd-v68dl_17d5b752-8e1d-4fe0-ad28-fc3073f96157/barbican-api/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.032924 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6658bb6dd-v68dl_17d5b752-8e1d-4fe0-ad28-fc3073f96157/barbican-api-log/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.254287 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f89475ccd-9rw9s_eb6c5d2c-fd76-44e2-a563-09b7dab7c352/barbican-keystone-listener/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.441449 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f89475ccd-9rw9s_eb6c5d2c-fd76-44e2-a563-09b7dab7c352/barbican-keystone-listener-log/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.584020 4754 generic.go:334] "Generic (PLEG): container finished" podID="937fe477-f212-4b46-9a9a-f4392418ffdc" containerID="0e58ff59d7027a106ed5101b0b6df53f7cc2c494bdbcfce02307097f0c0b6ba5" exitCode=0 Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.584085 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" event={"ID":"937fe477-f212-4b46-9a9a-f4392418ffdc","Type":"ContainerDied","Data":"0e58ff59d7027a106ed5101b0b6df53f7cc2c494bdbcfce02307097f0c0b6ba5"} Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.623383 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-xgqhb"] Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.637063 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xbq7q/crc-debug-xgqhb"] Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.645699 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-66d4bc859f-nl9q7_223cdab5-c653-4b55-a14f-c6206ee53f32/barbican-worker/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.728310 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-66d4bc859f-nl9q7_223cdab5-c653-4b55-a14f-c6206ee53f32/barbican-worker-log/0.log" Oct 11 04:30:48 crc kubenswrapper[4754]: I1011 04:30:48.949816 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-w6gvw_f77837ec-b2bb-4683-bc5f-21ec5145910c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.122807 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8244086c-5d5e-489a-b798-ed127ec9d89a/ceilometer-central-agent/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.203577 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8244086c-5d5e-489a-b798-ed127ec9d89a/ceilometer-notification-agent/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.235331 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8244086c-5d5e-489a-b798-ed127ec9d89a/proxy-httpd/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.313465 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8244086c-5d5e-489a-b798-ed127ec9d89a/sg-core/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.466161 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-rv8bz_986f6f39-c174-4f1e-a798-6ee6c389245e/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.750682 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.856098 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7n6f\" (UniqueName: \"kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f\") pod \"937fe477-f212-4b46-9a9a-f4392418ffdc\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.856208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host\") pod \"937fe477-f212-4b46-9a9a-f4392418ffdc\" (UID: \"937fe477-f212-4b46-9a9a-f4392418ffdc\") " Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.856710 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host" (OuterVolumeSpecName: "host") pod "937fe477-f212-4b46-9a9a-f4392418ffdc" (UID: "937fe477-f212-4b46-9a9a-f4392418ffdc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.868220 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f" (OuterVolumeSpecName: "kube-api-access-r7n6f") pod "937fe477-f212-4b46-9a9a-f4392418ffdc" (UID: "937fe477-f212-4b46-9a9a-f4392418ffdc"). InnerVolumeSpecName "kube-api-access-r7n6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.936643 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-t5gvv_72c8febd-5b78-489c-9fdd-3235935e9934/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.958355 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/937fe477-f212-4b46-9a9a-f4392418ffdc-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:49 crc kubenswrapper[4754]: I1011 04:30:49.958396 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7n6f\" (UniqueName: \"kubernetes.io/projected/937fe477-f212-4b46-9a9a-f4392418ffdc-kube-api-access-r7n6f\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.457553 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_61b305b9-2dcf-4961-a33d-d93a242b01cf/probe/0.log" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.612201 4754 scope.go:117] "RemoveContainer" containerID="0e58ff59d7027a106ed5101b0b6df53f7cc2c494bdbcfce02307097f0c0b6ba5" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.612248 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/crc-debug-xgqhb" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.842325 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_be981514-5d9c-4c0a-8853-12d2d04e6e76/cinder-api/0.log" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.948036 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5543cbf6-34ea-4bdc-86c3-7b61d6a14c03/cinder-scheduler/0.log" Oct 11 04:30:50 crc kubenswrapper[4754]: I1011 04:30:50.999835 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_be981514-5d9c-4c0a-8853-12d2d04e6e76/cinder-api-log/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.093945 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5543cbf6-34ea-4bdc-86c3-7b61d6a14c03/probe/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.101069 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="937fe477-f212-4b46-9a9a-f4392418ffdc" path="/var/lib/kubelet/pods/937fe477-f212-4b46-9a9a-f4392418ffdc/volumes" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.367356 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_ae6da214-d13e-4fde-9e99-1e8db2f6e248/probe/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.575236 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-bds68_dd2df864-0dde-4eaa-88f4-ae2e093087ec/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.661019 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_61b305b9-2dcf-4961-a33d-d93a242b01cf/cinder-backup/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.776800 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-qk6ps_1d13d97f-00e5-4264-8817-6ae83b60819c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:51 crc kubenswrapper[4754]: I1011 04:30:51.860391 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pz4xn_f3601d47-a321-4202-9813-6b9ce785cbee/init/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.157946 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pz4xn_f3601d47-a321-4202-9813-6b9ce785cbee/init/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.248547 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pz4xn_f3601d47-a321-4202-9813-6b9ce785cbee/dnsmasq-dns/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.373271 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca/glance-httpd/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.478001 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d0b8f6ee-2e89-4095-8e34-304a9cf7f8ca/glance-log/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.614946 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c8cc0a22-3499-4f27-9250-70c5a233a95f/glance-httpd/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.679591 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c8cc0a22-3499-4f27-9250-70c5a233a95f/glance-log/0.log" Oct 11 04:30:52 crc kubenswrapper[4754]: I1011 04:30:52.945566 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8655c657c6-4mw6n_96d0a63b-15e8-4830-8fd6-1bd47e6ed973/horizon/0.log" Oct 11 04:30:53 crc kubenswrapper[4754]: I1011 04:30:53.162329 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-mcfk8_4533b55f-2b04-4059-9ed9-f5be54f606b3/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:53 crc kubenswrapper[4754]: I1011 04:30:53.361590 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8655c657c6-4mw6n_96d0a63b-15e8-4830-8fd6-1bd47e6ed973/horizon-log/0.log" Oct 11 04:30:53 crc kubenswrapper[4754]: I1011 04:30:53.730739 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hjzrj_9a11ae05-4ae6-48a9-980d-84a802d36fb5/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.018401 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29335921-n6ljl_7d221e22-af78-451c-ae16-7650053c7b4d/keystone-cron/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.300005 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_955e7c85-4242-453b-b23d-432057450e28/kube-state-metrics/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.665069 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-748c895d7d-pnbk5_5ab66911-8a46-403c-9a45-e70b9cc46b60/keystone-api/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.686115 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-b59cc_347dff33-1661-4ed0-97b0-4248e8648f6a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.940094 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0aeef235-4c59-4b53-bb5a-5284a5c3066b/manila-api-log/0.log" Oct 11 04:30:54 crc kubenswrapper[4754]: I1011 04:30:54.995063 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0aeef235-4c59-4b53-bb5a-5284a5c3066b/manila-api/0.log" Oct 11 04:30:55 crc kubenswrapper[4754]: I1011 04:30:55.476741 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ce406d24-4a3d-4ee8-8793-b61a0a3e6dde/probe/0.log" Oct 11 04:30:55 crc kubenswrapper[4754]: I1011 04:30:55.511192 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ce406d24-4a3d-4ee8-8793-b61a0a3e6dde/manila-scheduler/0.log" Oct 11 04:30:55 crc kubenswrapper[4754]: I1011 04:30:55.740002 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_593cef67-51db-4722-a6c9-886fa38cc451/probe/0.log" Oct 11 04:30:55 crc kubenswrapper[4754]: I1011 04:30:55.801990 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_593cef67-51db-4722-a6c9-886fa38cc451/manila-share/0.log" Oct 11 04:30:56 crc kubenswrapper[4754]: I1011 04:30:56.503951 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cbfbbbc9-kgfv2_50322a6c-d0d3-4574-a584-e94ef7b39e70/neutron-api/0.log" Oct 11 04:30:56 crc kubenswrapper[4754]: I1011 04:30:56.515506 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cbfbbbc9-kgfv2_50322a6c-d0d3-4574-a584-e94ef7b39e70/neutron-httpd/0.log" Oct 11 04:30:56 crc kubenswrapper[4754]: I1011 04:30:56.731394 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4w2zg_e12f53c3-0230-41c0-a33a-d59605640584/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:57 crc kubenswrapper[4754]: I1011 04:30:57.753004 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b/nova-api-log/0.log" Oct 11 04:30:58 crc kubenswrapper[4754]: I1011 04:30:58.083349 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_21f0eb8a-e4a4-4415-9bca-5dbc66f8c16b/nova-api-api/0.log" Oct 11 04:30:58 crc kubenswrapper[4754]: I1011 04:30:58.359988 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_198c0494-b8ae-45a5-a666-b60c0de08d0f/nova-cell0-conductor-conductor/0.log" Oct 11 04:30:58 crc kubenswrapper[4754]: I1011 04:30:58.698584 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4401efc4-ecf2-42cd-b406-38206593f37e/nova-cell1-conductor-conductor/0.log" Oct 11 04:30:58 crc kubenswrapper[4754]: I1011 04:30:58.875651 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_dd32e65c-d404-481f-aa9a-18e3a0819262/nova-cell1-novncproxy-novncproxy/0.log" Oct 11 04:30:59 crc kubenswrapper[4754]: I1011 04:30:59.212809 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_ae6da214-d13e-4fde-9e99-1e8db2f6e248/cinder-volume/0.log" Oct 11 04:30:59 crc kubenswrapper[4754]: I1011 04:30:59.371607 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7htzb_ff20e060-f35b-4a0f-8861-a7befb21032e/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:30:59 crc kubenswrapper[4754]: I1011 04:30:59.502767 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_cd491934-e354-499d-ad2d-f77049e8e4b7/nova-metadata-log/0.log" Oct 11 04:30:59 crc kubenswrapper[4754]: I1011 04:30:59.947822 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d6f98545-924e-4087-8f8e-15c9f9b260cb/nova-scheduler-scheduler/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.129952 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01de3ace-e995-419f-ad27-75ee2eb0e110/mysql-bootstrap/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.387537 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01de3ace-e995-419f-ad27-75ee2eb0e110/mysql-bootstrap/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.490059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01de3ace-e995-419f-ad27-75ee2eb0e110/galera/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.717087 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6a40ecec-4fe2-41e1-8d6c-43e7689eedfe/mysql-bootstrap/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.962460 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6a40ecec-4fe2-41e1-8d6c-43e7689eedfe/mysql-bootstrap/0.log" Oct 11 04:31:00 crc kubenswrapper[4754]: I1011 04:31:00.976699 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6a40ecec-4fe2-41e1-8d6c-43e7689eedfe/galera/0.log" Oct 11 04:31:01 crc kubenswrapper[4754]: I1011 04:31:01.213955 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_89166926-b13c-442e-ba5f-b15200cc237a/openstackclient/0.log" Oct 11 04:31:01 crc kubenswrapper[4754]: I1011 04:31:01.467789 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7scx8_1842e17e-2c29-46ec-bdec-a6802d656e10/ovn-controller/0.log" Oct 11 04:31:01 crc kubenswrapper[4754]: I1011 04:31:01.642745 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_cd491934-e354-499d-ad2d-f77049e8e4b7/nova-metadata-metadata/0.log" Oct 11 04:31:01 crc kubenswrapper[4754]: I1011 04:31:01.652462 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m7h7p_406e66d8-4b00-4279-8acc-3e39d25b75c3/openstack-network-exporter/0.log" Oct 11 04:31:01 crc kubenswrapper[4754]: I1011 04:31:01.867849 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7sm4b_953544e1-701d-4d5c-831e-77e952fc8504/ovsdb-server-init/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.037013 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7sm4b_953544e1-701d-4d5c-831e-77e952fc8504/ovsdb-server-init/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.068666 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7sm4b_953544e1-701d-4d5c-831e-77e952fc8504/ovsdb-server/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.100252 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7sm4b_953544e1-701d-4d5c-831e-77e952fc8504/ovs-vswitchd/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.278056 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hxntc_51e593be-a82a-499d-91fd-f2e05d771519/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.478507 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3b76ddc-70a5-4142-8b7c-1342a716436f/openstack-network-exporter/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.483951 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3b76ddc-70a5-4142-8b7c-1342a716436f/ovn-northd/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.680019 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5fa3e4c1-2c10-4fd7-950f-380c41b8af4b/openstack-network-exporter/0.log" Oct 11 04:31:02 crc kubenswrapper[4754]: I1011 04:31:02.696136 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5fa3e4c1-2c10-4fd7-950f-380c41b8af4b/ovsdbserver-nb/0.log" Oct 11 04:31:03 crc kubenswrapper[4754]: I1011 04:31:03.376518 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_daa384d9-ff26-4962-872c-68d7ffea94fe/openstack-network-exporter/0.log" Oct 11 04:31:03 crc kubenswrapper[4754]: I1011 04:31:03.427769 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_daa384d9-ff26-4962-872c-68d7ffea94fe/ovsdbserver-sb/0.log" Oct 11 04:31:03 crc kubenswrapper[4754]: I1011 04:31:03.841236 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b68f46cdd-c77ph_218e1b80-55c5-427d-85ef-b35d891f69a0/placement-api/0.log" Oct 11 04:31:03 crc kubenswrapper[4754]: I1011 04:31:03.879454 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b68f46cdd-c77ph_218e1b80-55c5-427d-85ef-b35d891f69a0/placement-log/0.log" Oct 11 04:31:04 crc kubenswrapper[4754]: I1011 04:31:04.078722 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3b45e69c-6b94-4c29-8dcf-7a10c16a51be/setup-container/0.log" Oct 11 04:31:04 crc kubenswrapper[4754]: I1011 04:31:04.261952 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3b45e69c-6b94-4c29-8dcf-7a10c16a51be/setup-container/0.log" Oct 11 04:31:04 crc kubenswrapper[4754]: I1011 04:31:04.326695 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3b45e69c-6b94-4c29-8dcf-7a10c16a51be/rabbitmq/0.log" Oct 11 04:31:04 crc kubenswrapper[4754]: I1011 04:31:04.472704 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_aff682f3-e7eb-4034-8e82-27a6dd09490a/setup-container/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.037336 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_aff682f3-e7eb-4034-8e82-27a6dd09490a/setup-container/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.066737 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_aff682f3-e7eb-4034-8e82-27a6dd09490a/rabbitmq/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.230895 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kmrfg_f8f433cb-12ea-454a-8f71-61d75926dcc2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.379876 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-txhk4_11b8f98b-8629-4ebd-b448-89d3056021cf/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.572085 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rsgf8_e0749e96-0b16-40c1-9f1c-18227b93168b/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.728848 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-89n62_511691f1-e3fa-4f9f-9a58-8a095ef1d39c/ssh-known-hosts-edpm-deployment/0.log" Oct 11 04:31:05 crc kubenswrapper[4754]: I1011 04:31:05.950551 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1b1a049c-cb14-4ec6-8843-69950502bce7/tempest-tests-tempest-tests-runner/0.log" Oct 11 04:31:06 crc kubenswrapper[4754]: I1011 04:31:06.060745 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_8d37a1cd-4bdf-4c04-9347-40fce1afe82f/test-operator-logs-container/0.log" Oct 11 04:31:06 crc kubenswrapper[4754]: I1011 04:31:06.245815 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4wllv_4e956480-ea42-4f8b-96ce-cebce380af8a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 11 04:31:13 crc kubenswrapper[4754]: I1011 04:31:13.269264 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0dc66be1-e62f-4919-8124-9b571ac1849b/memcached/0.log" Oct 11 04:31:29 crc kubenswrapper[4754]: I1011 04:31:29.820791 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/util/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.009356 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/util/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.011206 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/pull/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.031355 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/pull/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.188892 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/pull/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.197024 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/util/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.198416 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2038d8801fe6c9f1493ccfe75e7dc3ce3601d64c42d7a5932fd9a792db7nchw_c3f7cf56-ada2-4541-a04a-bf1d56775252/extract/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.407933 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-54874b48b8-8vcz8_aae84d6e-3560-4f7b-85c4-bd3e34e98cae/kube-rbac-proxy/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.409433 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-6b4pp_303c4a96-7b77-49b6-8e02-8a23fefaae79/kube-rbac-proxy/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.430328 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-6b4pp_303c4a96-7b77-49b6-8e02-8a23fefaae79/manager/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.592489 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-5fgfl_64bb612c-78b4-4602-9531-358ffc64b40c/kube-rbac-proxy/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.615407 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-5fgfl_64bb612c-78b4-4602-9531-358ffc64b40c/manager/0.log" Oct 11 04:31:30 crc kubenswrapper[4754]: I1011 04:31:30.636954 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-54874b48b8-8vcz8_aae84d6e-3560-4f7b-85c4-bd3e34e98cae/manager/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.258026 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-4fftb_bbeae55c-0565-42d4-877e-2bcc53c30440/kube-rbac-proxy/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.335313 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-4fftb_bbeae55c-0565-42d4-877e-2bcc53c30440/manager/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.472299 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mbv4c_9db99c94-e7a3-4762-a719-cd7f92b477a6/kube-rbac-proxy/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.505851 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mbv4c_9db99c94-e7a3-4762-a719-cd7f92b477a6/manager/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.506835 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-dk7sg_c25997fd-2525-4846-a923-38d089a829d7/kube-rbac-proxy/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.676799 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-dk7sg_c25997fd-2525-4846-a923-38d089a829d7/manager/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.771316 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-ggjw9_d4d92b82-f730-4424-984c-0d1ec138e76c/kube-rbac-proxy/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.935201 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-26cn2_ebbf1f59-dc6c-4ebf-99a4-c9f806dce032/kube-rbac-proxy/0.log" Oct 11 04:31:31 crc kubenswrapper[4754]: I1011 04:31:31.947147 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-ggjw9_d4d92b82-f730-4424-984c-0d1ec138e76c/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.030692 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-26cn2_ebbf1f59-dc6c-4ebf-99a4-c9f806dce032/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.169220 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-dx268_64a954a8-d5af-48b6-961a-158e5fb739b7/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.183721 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-dx268_64a954a8-d5af-48b6-961a-158e5fb739b7/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.307161 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-h7hgd_5c1d2179-2db4-4e18-9f28-f8fcbd66cdec/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.408639 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-h7hgd_5c1d2179-2db4-4e18-9f28-f8fcbd66cdec/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.521263 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-t2725_7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.575771 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-t2725_7b7b8e95-80a6-49c2-8ab5-f8dddbe5c34b/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.600172 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-hjcs8_5f0ffe2a-9b82-443b-89d3-be54bdb596b6/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.775133 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-hjcs8_5f0ffe2a-9b82-443b-89d3-be54bdb596b6/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.784564 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7rvpp_1ceda0b1-1af1-4fe9-b522-02d9d9798556/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.937818 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7rvpp_1ceda0b1-1af1-4fe9-b522-02d9d9798556/manager/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.971733 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-k92pr_46931d30-913d-4738-86af-3fd45616e694/kube-rbac-proxy/0.log" Oct 11 04:31:32 crc kubenswrapper[4754]: I1011 04:31:32.972864 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-k92pr_46931d30-913d-4738-86af-3fd45616e694/manager/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.139526 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48_3654956b-08a4-4add-8706-9a5ab944c7f2/kube-rbac-proxy/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.216382 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dvwv48_3654956b-08a4-4add-8706-9a5ab944c7f2/manager/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.219630 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-594bbd78-2lt5n_35cfbe55-6993-4bb0-91d9-a05f8bfad5d4/kube-rbac-proxy/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.405054 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7c6467487b-bgwp8_3ec784c6-1183-4341-8d58-6cdd11b0b854/kube-rbac-proxy/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.633989 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-52qtt_2161836a-986e-4ed4-91ed-9132f0582a85/registry-server/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.642513 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7c6467487b-bgwp8_3ec784c6-1183-4341-8d58-6cdd11b0b854/operator/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.663326 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-h6mmf_e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e/kube-rbac-proxy/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.866411 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vxkjq_3bf92f20-fa39-4e8e-8092-f41f691aab74/kube-rbac-proxy/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.872084 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-h6mmf_e2f5b6be-db46-40a9-88f6-5a1fb0d8b30e/manager/0.log" Oct 11 04:31:33 crc kubenswrapper[4754]: I1011 04:31:33.974551 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vxkjq_3bf92f20-fa39-4e8e-8092-f41f691aab74/manager/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.185430 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nzvxp_2fe0ff99-f1c3-4446-8d92-653e447b239a/kube-rbac-proxy/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.185929 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-l9gzm_67767e7e-3d7b-4125-8695-aac89dc68d7f/operator/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.336952 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nzvxp_2fe0ff99-f1c3-4446-8d92-653e447b239a/manager/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.426058 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-kv5hr_8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c/kube-rbac-proxy/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.492753 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-kv5hr_8e5dfa13-7d80-435e-aa41-a1a6b6a2ca3c/manager/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.552986 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-594bbd78-2lt5n_35cfbe55-6993-4bb0-91d9-a05f8bfad5d4/manager/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.626745 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-f5rxz_8b5b1dbd-7d2f-4716-8a59-354854b70195/kube-rbac-proxy/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.742272 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-f5rxz_8b5b1dbd-7d2f-4716-8a59-354854b70195/manager/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.758375 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-qm42g_446576e5-02c4-4417-bb50-644f527631dd/kube-rbac-proxy/0.log" Oct 11 04:31:34 crc kubenswrapper[4754]: I1011 04:31:34.770406 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-qm42g_446576e5-02c4-4417-bb50-644f527631dd/manager/0.log" Oct 11 04:31:49 crc kubenswrapper[4754]: I1011 04:31:49.678927 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xcwj8_552772dc-229e-447d-a387-84d36f22cfe3/control-plane-machine-set-operator/0.log" Oct 11 04:31:49 crc kubenswrapper[4754]: I1011 04:31:49.860150 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dhcrp_8e720967-ded2-443d-a9ae-dadbeebebdaa/kube-rbac-proxy/0.log" Oct 11 04:31:49 crc kubenswrapper[4754]: I1011 04:31:49.943143 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dhcrp_8e720967-ded2-443d-a9ae-dadbeebebdaa/machine-api-operator/0.log" Oct 11 04:32:02 crc kubenswrapper[4754]: I1011 04:32:02.498151 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6j2fj_04bdf1ce-f0a8-4b76-870b-b513547b391a/cert-manager-controller/0.log" Oct 11 04:32:02 crc kubenswrapper[4754]: I1011 04:32:02.662414 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-p6m9q_14fb2f1e-9c40-462d-bd5e-db8bdcd0db23/cert-manager-cainjector/0.log" Oct 11 04:32:02 crc kubenswrapper[4754]: I1011 04:32:02.756517 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-g6lpk_f4adad4f-1d97-4526-9d87-01552818cb64/cert-manager-webhook/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.217400 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-txjs2_bb182ceb-1372-432d-9809-c67c03eb4091/nmstate-console-plugin/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.278235 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7qn6k_2a01ddcb-c062-4d74-91e7-a6793e5b79cb/nmstate-handler/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.409787 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2s4dm_d39cf2da-c123-476f-abc2-acfe1da12347/kube-rbac-proxy/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.449658 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2s4dm_d39cf2da-c123-476f-abc2-acfe1da12347/nmstate-metrics/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.564150 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-tvh84_dd5cec53-cb56-472d-94c9-af438d539f17/nmstate-operator/0.log" Oct 11 04:32:15 crc kubenswrapper[4754]: I1011 04:32:15.628159 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-b4vlh_446f8e6b-697d-4bfa-8464-e8b4decf2a33/nmstate-webhook/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.123882 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-9s44v_fa5eef15-bdc7-497b-aae1-0a6d1f5d3885/kube-rbac-proxy/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.313368 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-9s44v_fa5eef15-bdc7-497b-aae1-0a6d1f5d3885/controller/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.412885 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-6rwvw_01ff4e21-2909-48d9-a3e6-dae8528a1658/frr-k8s-webhook-server/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.513207 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-frr-files/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.721641 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-frr-files/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.725710 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-metrics/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.741611 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.741711 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.762843 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-reloader/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.769483 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-reloader/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.972794 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-frr-files/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.996504 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-metrics/0.log" Oct 11 04:32:30 crc kubenswrapper[4754]: I1011 04:32:30.997031 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-reloader/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.015568 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-metrics/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.173352 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-frr-files/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.215171 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-reloader/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.215290 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/cp-metrics/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.224392 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/controller/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.391210 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/kube-rbac-proxy-frr/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.396063 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/frr-metrics/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.419814 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/kube-rbac-proxy/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.642799 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/reloader/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.674824 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-67f8c59cbf-xd9pm_37697b5b-13b1-40b7-96e6-4e5e21499bec/manager/0.log" Oct 11 04:32:31 crc kubenswrapper[4754]: I1011 04:32:31.899024 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84c7b6d8cc-b6cv4_ebdfdd95-930a-43be-b2c4-1981a60df09e/webhook-server/0.log" Oct 11 04:32:32 crc kubenswrapper[4754]: I1011 04:32:32.113145 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49m5p_70ca9b96-8096-4d8a-b1e6-f954eb7de59f/kube-rbac-proxy/0.log" Oct 11 04:32:32 crc kubenswrapper[4754]: I1011 04:32:32.546635 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49m5p_70ca9b96-8096-4d8a-b1e6-f954eb7de59f/speaker/0.log" Oct 11 04:32:33 crc kubenswrapper[4754]: I1011 04:32:33.051523 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wjqkr_02c9fc4b-bec8-4468-b120-08f4db18ed07/frr/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.148057 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/util/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.310419 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/util/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.329680 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/pull/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.349446 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/pull/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.492859 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/pull/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.500540 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/extract/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.511020 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lmxnr_25829584-c627-4829-933c-fa2420a920b7/util/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.651711 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-utilities/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.900664 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-utilities/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.902821 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-content/0.log" Oct 11 04:32:45 crc kubenswrapper[4754]: I1011 04:32:45.902980 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-content/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.054562 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-content/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.097438 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/extract-utilities/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.254788 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-utilities/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.699091 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-utilities/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.825548 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlhdm_0f1054ac-3a3f-45e0-b99d-56fc3ccd35a5/registry-server/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.864697 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-content/0.log" Oct 11 04:32:46 crc kubenswrapper[4754]: I1011 04:32:46.965057 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-content/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.094646 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-content/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.103708 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/extract-utilities/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.342792 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/util/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.571543 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/util/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.616669 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/pull/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.625774 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/pull/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.710316 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kwzzg_54a0b5ff-4b0d-4ea9-ad2f-6deb0041233d/registry-server/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.806696 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/pull/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.811232 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/extract/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.835270 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm7ngs_5f472e73-bd23-4ff5-97de-14d0596c1b0c/util/0.log" Oct 11 04:32:47 crc kubenswrapper[4754]: I1011 04:32:47.945859 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-46rsl_ed58c00a-acd3-4c68-847f-2c731a66b7bf/marketplace-operator/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.004462 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.174878 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.175185 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-content/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.179741 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-content/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.328944 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-content/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.337285 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.390781 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.524367 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zt9gg_da6b7d01-e8c6-401e-8a3c-0f9504c28b86/registry-server/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.588994 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-content/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.597979 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-content/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.624083 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.759204 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-utilities/0.log" Oct 11 04:32:48 crc kubenswrapper[4754]: I1011 04:32:48.769676 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/extract-content/0.log" Oct 11 04:32:49 crc kubenswrapper[4754]: I1011 04:32:49.231349 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qs976_35574e71-4d0b-469d-a35e-dfbfa02664a3/registry-server/0.log" Oct 11 04:33:00 crc kubenswrapper[4754]: I1011 04:33:00.736800 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:33:00 crc kubenswrapper[4754]: I1011 04:33:00.737461 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:33:30 crc kubenswrapper[4754]: I1011 04:33:30.736437 4754 patch_prober.go:28] interesting pod/machine-config-daemon-px2sf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:33:30 crc kubenswrapper[4754]: I1011 04:33:30.737071 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:33:30 crc kubenswrapper[4754]: I1011 04:33:30.737127 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" Oct 11 04:33:30 crc kubenswrapper[4754]: I1011 04:33:30.738081 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038"} pod="openshift-machine-config-operator/machine-config-daemon-px2sf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:33:30 crc kubenswrapper[4754]: I1011 04:33:30.738479 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerName="machine-config-daemon" containerID="cri-o://6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" gracePeriod=600 Oct 11 04:33:30 crc kubenswrapper[4754]: E1011 04:33:30.888761 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:33:31 crc kubenswrapper[4754]: I1011 04:33:31.042130 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" exitCode=0 Oct 11 04:33:31 crc kubenswrapper[4754]: I1011 04:33:31.042174 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" event={"ID":"ef866bf9-b2ed-4ab9-922b-68f7373b6c04","Type":"ContainerDied","Data":"6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038"} Oct 11 04:33:31 crc kubenswrapper[4754]: I1011 04:33:31.042207 4754 scope.go:117] "RemoveContainer" containerID="0128e9ca94bdd46ff3ad83fd01b8dd34eb918451b0102b0a6a4d4dd3b72b1486" Oct 11 04:33:31 crc kubenswrapper[4754]: I1011 04:33:31.042948 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:33:31 crc kubenswrapper[4754]: E1011 04:33:31.043391 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.642932 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:43 crc kubenswrapper[4754]: E1011 04:33:43.643761 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="937fe477-f212-4b46-9a9a-f4392418ffdc" containerName="container-00" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.643774 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="937fe477-f212-4b46-9a9a-f4392418ffdc" containerName="container-00" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.650056 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="937fe477-f212-4b46-9a9a-f4392418ffdc" containerName="container-00" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.660967 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.696623 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.777555 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.777763 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.777816 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m86px\" (UniqueName: \"kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.829207 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.831462 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.839009 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.882501 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.882580 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m86px\" (UniqueName: \"kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.882672 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.883159 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.883388 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.984018 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqt7q\" (UniqueName: \"kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.984459 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:43 crc kubenswrapper[4754]: I1011 04:33:43.984502 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.083918 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:33:44 crc kubenswrapper[4754]: E1011 04:33:44.084198 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.086854 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqt7q\" (UniqueName: \"kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.087012 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.087062 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.087497 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.087531 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.171881 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m86px\" (UniqueName: \"kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px\") pod \"community-operators-mddd9\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.172458 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqt7q\" (UniqueName: \"kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q\") pod \"certified-operators-nm99w\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.300161 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.447129 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:44 crc kubenswrapper[4754]: I1011 04:33:44.876744 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:44 crc kubenswrapper[4754]: W1011 04:33:44.877999 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd98fdcf_c01d_4c67_bd71_f9bf5767acf3.slice/crio-51860a907b9416fc46e434793129a5e7affde6252dffcedd0eaaaf85e10e541c WatchSource:0}: Error finding container 51860a907b9416fc46e434793129a5e7affde6252dffcedd0eaaaf85e10e541c: Status 404 returned error can't find the container with id 51860a907b9416fc46e434793129a5e7affde6252dffcedd0eaaaf85e10e541c Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.017430 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:45 crc kubenswrapper[4754]: W1011 04:33:45.027751 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd88a62b3_1b4a_437e_bf02_484a2cb64d16.slice/crio-7743a6f36008ce77f401595474aa1e5132689e57266b9c5c6d9ca240945ec134 WatchSource:0}: Error finding container 7743a6f36008ce77f401595474aa1e5132689e57266b9c5c6d9ca240945ec134: Status 404 returned error can't find the container with id 7743a6f36008ce77f401595474aa1e5132689e57266b9c5c6d9ca240945ec134 Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.190211 4754 generic.go:334] "Generic (PLEG): container finished" podID="cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" containerID="0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613" exitCode=0 Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.190286 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerDied","Data":"0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613"} Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.190315 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerStarted","Data":"51860a907b9416fc46e434793129a5e7affde6252dffcedd0eaaaf85e10e541c"} Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.193419 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:33:45 crc kubenswrapper[4754]: I1011 04:33:45.198240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerStarted","Data":"7743a6f36008ce77f401595474aa1e5132689e57266b9c5c6d9ca240945ec134"} Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.209715 4754 generic.go:334] "Generic (PLEG): container finished" podID="d88a62b3-1b4a-437e-bf02-484a2cb64d16" containerID="0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f" exitCode=0 Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.209928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerDied","Data":"0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f"} Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.260625 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.263111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.305683 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.336891 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.337005 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xngf\" (UniqueName: \"kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.339901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.441815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.441943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.442016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xngf\" (UniqueName: \"kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.442727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.442873 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.466075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xngf\" (UniqueName: \"kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf\") pod \"redhat-operators-ldjxc\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:46 crc kubenswrapper[4754]: I1011 04:33:46.588103 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:47 crc kubenswrapper[4754]: I1011 04:33:47.101050 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:33:47 crc kubenswrapper[4754]: I1011 04:33:47.219565 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerStarted","Data":"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86"} Oct 11 04:33:47 crc kubenswrapper[4754]: I1011 04:33:47.222311 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerStarted","Data":"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a"} Oct 11 04:33:47 crc kubenswrapper[4754]: I1011 04:33:47.224508 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerStarted","Data":"6aa31d73a31a786e67efc195710449cbb5698c526c697b611fe2a7af3cc38dae"} Oct 11 04:33:48 crc kubenswrapper[4754]: I1011 04:33:48.235126 4754 generic.go:334] "Generic (PLEG): container finished" podID="cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" containerID="5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a" exitCode=0 Oct 11 04:33:48 crc kubenswrapper[4754]: I1011 04:33:48.236614 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerDied","Data":"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a"} Oct 11 04:33:48 crc kubenswrapper[4754]: I1011 04:33:48.240416 4754 generic.go:334] "Generic (PLEG): container finished" podID="2e28defb-65bd-412d-88b8-282b5b4a3323" containerID="36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441" exitCode=0 Oct 11 04:33:48 crc kubenswrapper[4754]: I1011 04:33:48.243854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerDied","Data":"36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441"} Oct 11 04:33:49 crc kubenswrapper[4754]: I1011 04:33:49.250649 4754 generic.go:334] "Generic (PLEG): container finished" podID="d88a62b3-1b4a-437e-bf02-484a2cb64d16" containerID="9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86" exitCode=0 Oct 11 04:33:49 crc kubenswrapper[4754]: I1011 04:33:49.250717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerDied","Data":"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86"} Oct 11 04:33:49 crc kubenswrapper[4754]: I1011 04:33:49.253246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerStarted","Data":"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882"} Oct 11 04:33:49 crc kubenswrapper[4754]: I1011 04:33:49.305122 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mddd9" podStartSLOduration=2.837661614 podStartE2EDuration="6.305103057s" podCreationTimestamp="2025-10-11 04:33:43 +0000 UTC" firstStartedPulling="2025-10-11 04:33:45.193152761 +0000 UTC m=+5272.752097546" lastFinishedPulling="2025-10-11 04:33:48.660594204 +0000 UTC m=+5276.219538989" observedRunningTime="2025-10-11 04:33:49.298012848 +0000 UTC m=+5276.856957643" watchObservedRunningTime="2025-10-11 04:33:49.305103057 +0000 UTC m=+5276.864047842" Oct 11 04:33:50 crc kubenswrapper[4754]: I1011 04:33:50.266001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerStarted","Data":"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda"} Oct 11 04:33:50 crc kubenswrapper[4754]: I1011 04:33:50.272826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerStarted","Data":"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765"} Oct 11 04:33:50 crc kubenswrapper[4754]: I1011 04:33:50.299753 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nm99w" podStartSLOduration=3.799783508 podStartE2EDuration="7.299729443s" podCreationTimestamp="2025-10-11 04:33:43 +0000 UTC" firstStartedPulling="2025-10-11 04:33:46.213456369 +0000 UTC m=+5273.772401154" lastFinishedPulling="2025-10-11 04:33:49.713402264 +0000 UTC m=+5277.272347089" observedRunningTime="2025-10-11 04:33:50.294824525 +0000 UTC m=+5277.853769300" watchObservedRunningTime="2025-10-11 04:33:50.299729443 +0000 UTC m=+5277.858674218" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.301449 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.303628 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.311225 4754 generic.go:334] "Generic (PLEG): container finished" podID="2e28defb-65bd-412d-88b8-282b5b4a3323" containerID="5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765" exitCode=0 Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.311263 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerDied","Data":"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765"} Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.448483 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.448538 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.912007 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:54 crc kubenswrapper[4754]: I1011 04:33:54.922178 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:55 crc kubenswrapper[4754]: I1011 04:33:55.323216 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerStarted","Data":"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101"} Oct 11 04:33:55 crc kubenswrapper[4754]: I1011 04:33:55.342275 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ldjxc" podStartSLOduration=2.819872198 podStartE2EDuration="9.342258167s" podCreationTimestamp="2025-10-11 04:33:46 +0000 UTC" firstStartedPulling="2025-10-11 04:33:48.248109378 +0000 UTC m=+5275.807054173" lastFinishedPulling="2025-10-11 04:33:54.770495347 +0000 UTC m=+5282.329440142" observedRunningTime="2025-10-11 04:33:55.340105737 +0000 UTC m=+5282.899050542" watchObservedRunningTime="2025-10-11 04:33:55.342258167 +0000 UTC m=+5282.901202942" Oct 11 04:33:55 crc kubenswrapper[4754]: I1011 04:33:55.372734 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:55 crc kubenswrapper[4754]: I1011 04:33:55.373742 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:56 crc kubenswrapper[4754]: I1011 04:33:56.083228 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:33:56 crc kubenswrapper[4754]: E1011 04:33:56.083779 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:33:56 crc kubenswrapper[4754]: I1011 04:33:56.589238 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:56 crc kubenswrapper[4754]: I1011 04:33:56.589350 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.216938 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.341356 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nm99w" podUID="d88a62b3-1b4a-437e-bf02-484a2cb64d16" containerName="registry-server" containerID="cri-o://7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda" gracePeriod=2 Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.712833 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ldjxc" podUID="2e28defb-65bd-412d-88b8-282b5b4a3323" containerName="registry-server" probeResult="failure" output=< Oct 11 04:33:57 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 11 04:33:57 crc kubenswrapper[4754]: > Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.877871 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.915068 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content\") pod \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.915124 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities\") pod \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.915225 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqt7q\" (UniqueName: \"kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q\") pod \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\" (UID: \"d88a62b3-1b4a-437e-bf02-484a2cb64d16\") " Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.918883 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities" (OuterVolumeSpecName: "utilities") pod "d88a62b3-1b4a-437e-bf02-484a2cb64d16" (UID: "d88a62b3-1b4a-437e-bf02-484a2cb64d16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.926391 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q" (OuterVolumeSpecName: "kube-api-access-qqt7q") pod "d88a62b3-1b4a-437e-bf02-484a2cb64d16" (UID: "d88a62b3-1b4a-437e-bf02-484a2cb64d16"). InnerVolumeSpecName "kube-api-access-qqt7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:33:57 crc kubenswrapper[4754]: I1011 04:33:57.966754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d88a62b3-1b4a-437e-bf02-484a2cb64d16" (UID: "d88a62b3-1b4a-437e-bf02-484a2cb64d16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.017553 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.017591 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88a62b3-1b4a-437e-bf02-484a2cb64d16-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.017607 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqt7q\" (UniqueName: \"kubernetes.io/projected/d88a62b3-1b4a-437e-bf02-484a2cb64d16-kube-api-access-qqt7q\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.217731 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.351785 4754 generic.go:334] "Generic (PLEG): container finished" podID="d88a62b3-1b4a-437e-bf02-484a2cb64d16" containerID="7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda" exitCode=0 Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.351880 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm99w" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.351924 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerDied","Data":"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda"} Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.351955 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm99w" event={"ID":"d88a62b3-1b4a-437e-bf02-484a2cb64d16","Type":"ContainerDied","Data":"7743a6f36008ce77f401595474aa1e5132689e57266b9c5c6d9ca240945ec134"} Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.351990 4754 scope.go:117] "RemoveContainer" containerID="7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.352027 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mddd9" podUID="cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" containerName="registry-server" containerID="cri-o://1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882" gracePeriod=2 Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.372276 4754 scope.go:117] "RemoveContainer" containerID="9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.390177 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.398835 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nm99w"] Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.405200 4754 scope.go:117] "RemoveContainer" containerID="0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.549554 4754 scope.go:117] "RemoveContainer" containerID="7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda" Oct 11 04:33:58 crc kubenswrapper[4754]: E1011 04:33:58.550123 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda\": container with ID starting with 7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda not found: ID does not exist" containerID="7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.550181 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda"} err="failed to get container status \"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda\": rpc error: code = NotFound desc = could not find container \"7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda\": container with ID starting with 7d7d3aa832a3f898f009c5b76f17b65a466075e2e98288393571a367c1253fda not found: ID does not exist" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.550213 4754 scope.go:117] "RemoveContainer" containerID="9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86" Oct 11 04:33:58 crc kubenswrapper[4754]: E1011 04:33:58.550586 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86\": container with ID starting with 9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86 not found: ID does not exist" containerID="9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.550699 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86"} err="failed to get container status \"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86\": rpc error: code = NotFound desc = could not find container \"9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86\": container with ID starting with 9274e7c1f713be94df9e282c3ed0d7be1f62bf12c726db9698e6b0b680410a86 not found: ID does not exist" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.550814 4754 scope.go:117] "RemoveContainer" containerID="0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f" Oct 11 04:33:58 crc kubenswrapper[4754]: E1011 04:33:58.551242 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f\": container with ID starting with 0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f not found: ID does not exist" containerID="0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.551274 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f"} err="failed to get container status \"0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f\": rpc error: code = NotFound desc = could not find container \"0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f\": container with ID starting with 0cfa9d3697f14ffef6ad9471feeb2cf3aa73bdbf834b69e9497f3453b2743a9f not found: ID does not exist" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.850684 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.935321 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities\") pod \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.935511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content\") pod \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.935748 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m86px\" (UniqueName: \"kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px\") pod \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\" (UID: \"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3\") " Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.936169 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities" (OuterVolumeSpecName: "utilities") pod "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" (UID: "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.936673 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:58 crc kubenswrapper[4754]: I1011 04:33:58.941754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px" (OuterVolumeSpecName: "kube-api-access-m86px") pod "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" (UID: "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3"). InnerVolumeSpecName "kube-api-access-m86px". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.002816 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" (UID: "cd98fdcf-c01d-4c67-bd71-f9bf5767acf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.039122 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.039177 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m86px\" (UniqueName: \"kubernetes.io/projected/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3-kube-api-access-m86px\") on node \"crc\" DevicePath \"\"" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.094643 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88a62b3-1b4a-437e-bf02-484a2cb64d16" path="/var/lib/kubelet/pods/d88a62b3-1b4a-437e-bf02-484a2cb64d16/volumes" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.364257 4754 generic.go:334] "Generic (PLEG): container finished" podID="cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" containerID="1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882" exitCode=0 Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.364325 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerDied","Data":"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882"} Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.364384 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mddd9" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.364411 4754 scope.go:117] "RemoveContainer" containerID="1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.364393 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mddd9" event={"ID":"cd98fdcf-c01d-4c67-bd71-f9bf5767acf3","Type":"ContainerDied","Data":"51860a907b9416fc46e434793129a5e7affde6252dffcedd0eaaaf85e10e541c"} Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.396095 4754 scope.go:117] "RemoveContainer" containerID="5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.398458 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.408610 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mddd9"] Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.418430 4754 scope.go:117] "RemoveContainer" containerID="0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.434100 4754 scope.go:117] "RemoveContainer" containerID="1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882" Oct 11 04:33:59 crc kubenswrapper[4754]: E1011 04:33:59.434621 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882\": container with ID starting with 1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882 not found: ID does not exist" containerID="1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.434717 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882"} err="failed to get container status \"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882\": rpc error: code = NotFound desc = could not find container \"1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882\": container with ID starting with 1898d14f5b1368dfbdf9592748c04ef7e69126bed0c6d60e5c71b16388c2e882 not found: ID does not exist" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.434750 4754 scope.go:117] "RemoveContainer" containerID="5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a" Oct 11 04:33:59 crc kubenswrapper[4754]: E1011 04:33:59.435943 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a\": container with ID starting with 5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a not found: ID does not exist" containerID="5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.436087 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a"} err="failed to get container status \"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a\": rpc error: code = NotFound desc = could not find container \"5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a\": container with ID starting with 5c3a5360393087f67fb6d8152332c52288292c4af0f522ca15a16dd0a7f55c7a not found: ID does not exist" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.436205 4754 scope.go:117] "RemoveContainer" containerID="0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613" Oct 11 04:33:59 crc kubenswrapper[4754]: E1011 04:33:59.436718 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613\": container with ID starting with 0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613 not found: ID does not exist" containerID="0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613" Oct 11 04:33:59 crc kubenswrapper[4754]: I1011 04:33:59.436746 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613"} err="failed to get container status \"0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613\": rpc error: code = NotFound desc = could not find container \"0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613\": container with ID starting with 0e22a9474a3da4bfcffd9c15b40c576b858f91e7c886891c5e05965a958bb613 not found: ID does not exist" Oct 11 04:34:01 crc kubenswrapper[4754]: I1011 04:34:01.095990 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd98fdcf-c01d-4c67-bd71-f9bf5767acf3" path="/var/lib/kubelet/pods/cd98fdcf-c01d-4c67-bd71-f9bf5767acf3/volumes" Oct 11 04:34:06 crc kubenswrapper[4754]: I1011 04:34:06.642306 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:34:06 crc kubenswrapper[4754]: I1011 04:34:06.692175 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:34:06 crc kubenswrapper[4754]: I1011 04:34:06.878274 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:34:08 crc kubenswrapper[4754]: I1011 04:34:08.464105 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ldjxc" podUID="2e28defb-65bd-412d-88b8-282b5b4a3323" containerName="registry-server" containerID="cri-o://821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101" gracePeriod=2 Oct 11 04:34:08 crc kubenswrapper[4754]: I1011 04:34:08.930732 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.042001 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities\") pod \"2e28defb-65bd-412d-88b8-282b5b4a3323\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.042241 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xngf\" (UniqueName: \"kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf\") pod \"2e28defb-65bd-412d-88b8-282b5b4a3323\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.042385 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content\") pod \"2e28defb-65bd-412d-88b8-282b5b4a3323\" (UID: \"2e28defb-65bd-412d-88b8-282b5b4a3323\") " Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.043242 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities" (OuterVolumeSpecName: "utilities") pod "2e28defb-65bd-412d-88b8-282b5b4a3323" (UID: "2e28defb-65bd-412d-88b8-282b5b4a3323"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.048370 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf" (OuterVolumeSpecName: "kube-api-access-4xngf") pod "2e28defb-65bd-412d-88b8-282b5b4a3323" (UID: "2e28defb-65bd-412d-88b8-282b5b4a3323"). InnerVolumeSpecName "kube-api-access-4xngf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.058105 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.058156 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xngf\" (UniqueName: \"kubernetes.io/projected/2e28defb-65bd-412d-88b8-282b5b4a3323-kube-api-access-4xngf\") on node \"crc\" DevicePath \"\"" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.150972 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e28defb-65bd-412d-88b8-282b5b4a3323" (UID: "2e28defb-65bd-412d-88b8-282b5b4a3323"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.160378 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e28defb-65bd-412d-88b8-282b5b4a3323-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.476260 4754 generic.go:334] "Generic (PLEG): container finished" podID="2e28defb-65bd-412d-88b8-282b5b4a3323" containerID="821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101" exitCode=0 Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.476327 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldjxc" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.476329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerDied","Data":"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101"} Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.476406 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldjxc" event={"ID":"2e28defb-65bd-412d-88b8-282b5b4a3323","Type":"ContainerDied","Data":"6aa31d73a31a786e67efc195710449cbb5698c526c697b611fe2a7af3cc38dae"} Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.476437 4754 scope.go:117] "RemoveContainer" containerID="821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.513829 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.519300 4754 scope.go:117] "RemoveContainer" containerID="5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.524650 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ldjxc"] Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.536310 4754 scope.go:117] "RemoveContainer" containerID="36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.577073 4754 scope.go:117] "RemoveContainer" containerID="821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101" Oct 11 04:34:09 crc kubenswrapper[4754]: E1011 04:34:09.577818 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101\": container with ID starting with 821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101 not found: ID does not exist" containerID="821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.578022 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101"} err="failed to get container status \"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101\": rpc error: code = NotFound desc = could not find container \"821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101\": container with ID starting with 821527beda3ced12bce15c9b3e3f528b6e4282a5395cd913bec6692f1fac8101 not found: ID does not exist" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.578064 4754 scope.go:117] "RemoveContainer" containerID="5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765" Oct 11 04:34:09 crc kubenswrapper[4754]: E1011 04:34:09.578355 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765\": container with ID starting with 5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765 not found: ID does not exist" containerID="5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.578383 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765"} err="failed to get container status \"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765\": rpc error: code = NotFound desc = could not find container \"5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765\": container with ID starting with 5c6f6169ecc87c36865e35392b07e2ccd7dd65098529f9a33d902852446da765 not found: ID does not exist" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.578399 4754 scope.go:117] "RemoveContainer" containerID="36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441" Oct 11 04:34:09 crc kubenswrapper[4754]: E1011 04:34:09.578634 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441\": container with ID starting with 36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441 not found: ID does not exist" containerID="36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441" Oct 11 04:34:09 crc kubenswrapper[4754]: I1011 04:34:09.578653 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441"} err="failed to get container status \"36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441\": rpc error: code = NotFound desc = could not find container \"36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441\": container with ID starting with 36253347edb45d94b95c4bdd369aea2545d6ad7209f7772f513871c074f54441 not found: ID does not exist" Oct 11 04:34:10 crc kubenswrapper[4754]: I1011 04:34:10.083371 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:34:10 crc kubenswrapper[4754]: E1011 04:34:10.083685 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:34:11 crc kubenswrapper[4754]: I1011 04:34:11.099536 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e28defb-65bd-412d-88b8-282b5b4a3323" path="/var/lib/kubelet/pods/2e28defb-65bd-412d-88b8-282b5b4a3323/volumes" Oct 11 04:34:22 crc kubenswrapper[4754]: I1011 04:34:22.083921 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:34:22 crc kubenswrapper[4754]: E1011 04:34:22.084769 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:34:34 crc kubenswrapper[4754]: I1011 04:34:34.084312 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:34:34 crc kubenswrapper[4754]: E1011 04:34:34.085599 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:34:46 crc kubenswrapper[4754]: I1011 04:34:46.083340 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:34:46 crc kubenswrapper[4754]: E1011 04:34:46.084053 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:34:46 crc kubenswrapper[4754]: I1011 04:34:46.842348 4754 generic.go:334] "Generic (PLEG): container finished" podID="da1fb582-ac79-4d1a-8380-2feb95343104" containerID="e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569" exitCode=0 Oct 11 04:34:46 crc kubenswrapper[4754]: I1011 04:34:46.842442 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" event={"ID":"da1fb582-ac79-4d1a-8380-2feb95343104","Type":"ContainerDied","Data":"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569"} Oct 11 04:34:46 crc kubenswrapper[4754]: I1011 04:34:46.843788 4754 scope.go:117] "RemoveContainer" containerID="e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569" Oct 11 04:34:47 crc kubenswrapper[4754]: I1011 04:34:47.164506 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xbq7q_must-gather-lfcgj_da1fb582-ac79-4d1a-8380-2feb95343104/gather/0.log" Oct 11 04:34:54 crc kubenswrapper[4754]: I1011 04:34:54.943020 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xbq7q/must-gather-lfcgj"] Oct 11 04:34:54 crc kubenswrapper[4754]: I1011 04:34:54.943784 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" podUID="da1fb582-ac79-4d1a-8380-2feb95343104" containerName="copy" containerID="cri-o://e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5" gracePeriod=2 Oct 11 04:34:54 crc kubenswrapper[4754]: I1011 04:34:54.955426 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xbq7q/must-gather-lfcgj"] Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.405231 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xbq7q_must-gather-lfcgj_da1fb582-ac79-4d1a-8380-2feb95343104/copy/0.log" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.405900 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.465100 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output\") pod \"da1fb582-ac79-4d1a-8380-2feb95343104\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.465244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4ct\" (UniqueName: \"kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct\") pod \"da1fb582-ac79-4d1a-8380-2feb95343104\" (UID: \"da1fb582-ac79-4d1a-8380-2feb95343104\") " Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.472679 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct" (OuterVolumeSpecName: "kube-api-access-qd4ct") pod "da1fb582-ac79-4d1a-8380-2feb95343104" (UID: "da1fb582-ac79-4d1a-8380-2feb95343104"). InnerVolumeSpecName "kube-api-access-qd4ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.568657 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4ct\" (UniqueName: \"kubernetes.io/projected/da1fb582-ac79-4d1a-8380-2feb95343104-kube-api-access-qd4ct\") on node \"crc\" DevicePath \"\"" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.621251 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "da1fb582-ac79-4d1a-8380-2feb95343104" (UID: "da1fb582-ac79-4d1a-8380-2feb95343104"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.670829 4754 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/da1fb582-ac79-4d1a-8380-2feb95343104-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.952721 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xbq7q_must-gather-lfcgj_da1fb582-ac79-4d1a-8380-2feb95343104/copy/0.log" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.953065 4754 generic.go:334] "Generic (PLEG): container finished" podID="da1fb582-ac79-4d1a-8380-2feb95343104" containerID="e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5" exitCode=143 Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.953123 4754 scope.go:117] "RemoveContainer" containerID="e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.953296 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xbq7q/must-gather-lfcgj" Oct 11 04:34:55 crc kubenswrapper[4754]: I1011 04:34:55.984353 4754 scope.go:117] "RemoveContainer" containerID="e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569" Oct 11 04:34:56 crc kubenswrapper[4754]: I1011 04:34:56.097888 4754 scope.go:117] "RemoveContainer" containerID="e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5" Oct 11 04:34:56 crc kubenswrapper[4754]: E1011 04:34:56.102121 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5\": container with ID starting with e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5 not found: ID does not exist" containerID="e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5" Oct 11 04:34:56 crc kubenswrapper[4754]: I1011 04:34:56.102243 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5"} err="failed to get container status \"e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5\": rpc error: code = NotFound desc = could not find container \"e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5\": container with ID starting with e8db7b423bdb2b2e2c8590ddab471c1813222af0482a265cd293c83da5238ed5 not found: ID does not exist" Oct 11 04:34:56 crc kubenswrapper[4754]: I1011 04:34:56.102328 4754 scope.go:117] "RemoveContainer" containerID="e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569" Oct 11 04:34:56 crc kubenswrapper[4754]: E1011 04:34:56.110216 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569\": container with ID starting with e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569 not found: ID does not exist" containerID="e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569" Oct 11 04:34:56 crc kubenswrapper[4754]: I1011 04:34:56.110424 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569"} err="failed to get container status \"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569\": rpc error: code = NotFound desc = could not find container \"e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569\": container with ID starting with e29609e3cff74d571128c0c495ff0e7d3640ce75a330e37c78452d21e4648569 not found: ID does not exist" Oct 11 04:34:57 crc kubenswrapper[4754]: I1011 04:34:57.094103 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da1fb582-ac79-4d1a-8380-2feb95343104" path="/var/lib/kubelet/pods/da1fb582-ac79-4d1a-8380-2feb95343104/volumes" Oct 11 04:34:58 crc kubenswrapper[4754]: I1011 04:34:58.087436 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:34:58 crc kubenswrapper[4754]: E1011 04:34:58.088714 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:35:11 crc kubenswrapper[4754]: I1011 04:35:11.107000 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:35:11 crc kubenswrapper[4754]: E1011 04:35:11.111685 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:35:24 crc kubenswrapper[4754]: I1011 04:35:24.084576 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:35:24 crc kubenswrapper[4754]: E1011 04:35:24.086053 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:35:35 crc kubenswrapper[4754]: I1011 04:35:35.083392 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:35:35 crc kubenswrapper[4754]: E1011 04:35:35.084324 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:35:48 crc kubenswrapper[4754]: I1011 04:35:48.085116 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:35:48 crc kubenswrapper[4754]: E1011 04:35:48.085861 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:36:01 crc kubenswrapper[4754]: I1011 04:36:01.083843 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:36:01 crc kubenswrapper[4754]: E1011 04:36:01.085123 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:36:13 crc kubenswrapper[4754]: I1011 04:36:13.090642 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:36:13 crc kubenswrapper[4754]: E1011 04:36:13.091313 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" Oct 11 04:36:28 crc kubenswrapper[4754]: I1011 04:36:28.084537 4754 scope.go:117] "RemoveContainer" containerID="6277f90af55e02af4a81fca1137c83aa230f225767174b8ab500d35963dd8038" Oct 11 04:36:28 crc kubenswrapper[4754]: E1011 04:36:28.085472 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-px2sf_openshift-machine-config-operator(ef866bf9-b2ed-4ab9-922b-68f7373b6c04)\"" pod="openshift-machine-config-operator/machine-config-daemon-px2sf" podUID="ef866bf9-b2ed-4ab9-922b-68f7373b6c04" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515072357325024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015072357325017373 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015072344307016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015072344307015462 5ustar corecore